diff --git a/LCS/PyCommon/dbcredentials.py b/LCS/PyCommon/dbcredentials.py
index 55e79a73cbc3900cb1932ae371697290914379fc..ee244daebb0c3f480629037c00e1b0cc58e6a356 100644
--- a/LCS/PyCommon/dbcredentials.py
+++ b/LCS/PyCommon/dbcredentials.py
@@ -69,10 +69,10 @@ class Credentials:
     self.config = {}
 
   def __str__(self):
-    return "type={type} addr={host}:{port} auth={user}:{password} db={database}".format(**self.__dict__)
+    return "db={database} addr={host}:{port} auth={user}:{password} type={type}".format(**self.__dict__)
 
   def stringWithHiddenPassword(self):
-    return "type={type} addr={host}:{port} auth={user}:XXXXXX db={database}".format(**self.__dict__)
+    return "db={database} addr={host}:{port} auth={user}:XXXXXX type={type}".format(**self.__dict__)
 
   def pg_connect_options(self):
     """
diff --git a/LCS/PyCommon/postgres.py b/LCS/PyCommon/postgres.py
index c43d56541bff70529b308720b8b72fd7197d72ff..13d6cee1db2a03e2b86863a3c08e174a2cc3ad01 100644
--- a/LCS/PyCommon/postgres.py
+++ b/LCS/PyCommon/postgres.py
@@ -116,6 +116,7 @@ class PostgresDatabaseConnection:
                  query_timeout: float=3600):
         self._dbcreds = dbcreds
         self._connection = None
+        self._cursor = None
         self.__auto_commit_selects = auto_commit_selects
         self.__num_connect_retries = num_connect_retries
         self.__connect_retry_interval = connect_retry_interval
@@ -132,7 +133,7 @@ class PostgresDatabaseConnection:
 
         for retry_cntr in range(self.__num_connect_retries+1):
             try:
-                logger.debug("trying to connect to database: %s", self)
+                logger.debug("connecting to database: %s", self)
 
                 self._connection = psycopg2.connect(host=self._dbcreds.host,
                                                     user=self._dbcreds.user,
@@ -172,13 +173,18 @@ class PostgresDatabaseConnection:
                     raise PostgresDBError(error_string)
 
     def disconnect(self):
-        if self.is_connected:
-            logger.debug("%s disconnecting from db: %s", self.__class__.__name__, self.database)
-            self._cursor.close()
-            self._cursor = None
-            self._connection.close()
-            self._connection = None
-            logger.debug("%s disconnected from db: %s", self.__class__.__name__, self.database)
+        if self._connection is not None or self._cursor is not None:
+            logger.debug("disconnecting from database: %s", self)
+
+            if self._cursor is not None:
+                self._cursor.close()
+                self._cursor = None
+
+            if self._connection is not None:
+                self._connection.close()
+                self._connection = None
+
+            logger.info("disconnected from database: %s", self)
 
     def _is_recoverable_connection_error(self, error: psycopg2.DatabaseError) -> bool:
         '''test if psycopg2.DatabaseError is a recoverable connection error'''
diff --git a/LCS/PyCommon/test/t_postgres.py b/LCS/PyCommon/test/t_postgres.py
index 7a0fdf1bfe5958a65488b26a657fc26ac01db114..29368d849be44e33dde4387c9b4ff9a0397b1fb9 100755
--- a/LCS/PyCommon/test/t_postgres.py
+++ b/LCS/PyCommon/test/t_postgres.py
@@ -65,7 +65,7 @@ class TestPostgres(unittest.TestCase):
 
             # check logging
             self.assertEqual(NUM_CONNECT_RETRIES, len([ca for ca in mocked_logger.info.call_args_list if 'retrying to connect' in ca[0][0]]))
-            self.assertEqual(NUM_CONNECT_RETRIES+1, len([ca for ca in mocked_logger.debug.call_args_list if 'trying to connect to database' in ca[0][0]]))
+            self.assertEqual(NUM_CONNECT_RETRIES+1, len([ca for ca in mocked_logger.debug.call_args_list if 'connecting to database' in ca[0][0]]))
             self.assertEqual(NUM_CONNECT_RETRIES+1, len([ca for ca in mocked_logger.error.call_args_list if 'could not connect' in ca[0][0]]))
 
     def test_reconnect_on_connection_loss(self):
diff --git a/MAC/Services/src/ObservationControl2.py b/MAC/Services/src/ObservationControl2.py
index 2c4854c7db5b8afb2e0bce490762772aeed72863..d3a10f64eac62d708bc20f5dcbfabcfa4f8e97f4 100644
--- a/MAC/Services/src/ObservationControl2.py
+++ b/MAC/Services/src/ObservationControl2.py
@@ -36,6 +36,7 @@ logger = logging.getLogger(__name__)
 class ObservationControlHandler(ServiceMessageHandler):
     def __init__(self):
         super(ObservationControlHandler, self).__init__()
+        self.register_service_method("AbortObservation", self.abort_observation)
 
         host = "localhost"
 
diff --git a/SAS/ResourceAssignment/Common/lib/specification.py b/SAS/ResourceAssignment/Common/lib/specification.py
index f3018278f3eb2ea056bff7853b722af28531ad96..f674547379642cee263a6270be966a008910278f 100644
--- a/SAS/ResourceAssignment/Common/lib/specification.py
+++ b/SAS/ResourceAssignment/Common/lib/specification.py
@@ -32,6 +32,8 @@ from lofar.parameterset import parameterset
 from datetime import datetime, timedelta
 from lofar.common.datetimeutils import parseDatetime
 from lofar.sas.resourceassignment.resourceassigner.schedulechecker import movePipelineAfterItsPredecessors
+from lofar.common.postgres import PostgresDBQueryExecutionError
+
 import pprint
 
 import logging
@@ -51,20 +53,20 @@ OUTPUT_PREFIX = "LOFAR.ObsSW."
 
 
 class Specification:
-    def __init__(self, otdbrpc, momquery, radbrpc):
+    def __init__(self, otdbrpc, momquery, radb):
         """Right now we internally represent a subset of a virtual instrument tree in OTDB and manipulate parset type
         specifications. Therefore we need an otdbid. At some point the OTDB dependency can hopefully be replaced with
         a proper virtual instrument model, when LOFAR is less dependent on MoM and OTDB and the Scheduler.
 
         :param otdbrpc: rpc to talk to otdb
         :param momquery:  rpc to query MoM
-        :param radbrpc: rpc to talk to radb
+        :param radb: RADB or RADBRPC instance
 
         In practice we need all 3 RPCs as we do consistency checking between the three databases for things like Cluster
         names.
         """
         self.otdbrpc  = otdbrpc
-        self.radbrpc = radbrpc
+        self.radb = radb
         self.momquery = momquery
 
         # I have not implemented getter/setter functions for most, maybe in a further refactor if it seems to be useful
@@ -166,28 +168,28 @@ class Specification:
         :param input_dict: Serialized version of a Specification and any predecessors.
         """
         self.otdb_id       = input_dict["otdb_id"]
-        self.mom_id        = input_dict["mom_id"]
+        self.mom_id        = input_dict.get("mom_id")
         self.radb_id       = input_dict["task_id"]
-        self.trigger_id    = input_dict["trigger_id"]
+        self.trigger_id    = input_dict.get("trigger_id")
         self.status        = input_dict["status"]
         self.type          = input_dict["task_type"]
-        self.subtype       = input_dict["task_subtype"]
-        self.starttime     = Specification.parse_datetime(input_dict["starttime"])
-        self.endtime       = Specification.parse_datetime(input_dict["endtime"])
-        self.duration      = Specification.parse_timedelta(input_dict["duration"])
-        self.min_starttime = Specification.parse_datetime(input_dict["min_starttime"])
-        self.max_endtime   = Specification.parse_datetime(input_dict["max_endtime"])
-        self.min_duration  = Specification.parse_timedelta(input_dict["min_duration"])
-        self.max_duration  = Specification.parse_timedelta(input_dict["max_duration"])
-        self.cluster       = input_dict["cluster"]
-        self.internal_dict = input_dict["specification"]
+        self.subtype       = input_dict.get("task_subtype")
+        self.starttime     = Specification.parse_datetime(input_dict.get("starttime"))
+        self.endtime       = Specification.parse_datetime(input_dict.get("endtime"))
+        self.duration      = Specification.parse_timedelta(input_dict.get("duration"))
+        self.min_starttime = Specification.parse_datetime(input_dict.get("min_starttime"))
+        self.max_endtime   = Specification.parse_datetime(input_dict.get("max_endtime"))
+        self.min_duration  = Specification.parse_timedelta(input_dict.get("min_duration"))
+        self.max_duration  = Specification.parse_timedelta(input_dict.get("max_duration"))
+        self.cluster       = input_dict.get("cluster")
+        self.internal_dict = input_dict.get("specification", {})
         self.predecessors  = []
-        for p in input_dict["predecessors"]:
-            spec = Specification(self.otdbrpc, self.momquery, self.radbrpc)
+        for p in input_dict.get("predecessors",[]):
+            spec = Specification(self.otdbrpc, self.momquery, self.radb)
             spec.from_dict(p)
             self.predecessors.append(spec)
-        self.successor_ids = input_dict["successors"]
-        self.storagemanager = input_dict["storagemanager"]
+        self.successor_ids = input_dict.get("successors",[])
+        self.storagemanager = input_dict.get("storagemanager")
 
     def isObservation(self):
         """:return if the Specification is for an observation."""
@@ -691,7 +693,7 @@ class Specification:
             if pred_otdb_id in found_specifications:
                 self.predecessors.append(found_specifications[pred_otdb_id])
             else:
-                spec = Specification(self.otdbrpc, self.momquery, self.radbrpc)
+                spec = Specification(self.otdbrpc, self.momquery, self.radb)
                 spec.read_from_OTDB_with_predecessors(pred_otdb_id, "otdb", found_specifications)
                 self.predecessors.append(spec) #TODO this needs a try/except somewhere?
         logger.info("Done reading from OTDB with predecessors")
@@ -716,7 +718,7 @@ class Specification:
             if str() in cluster_name_set or len(cluster_name_set) != 1:
                 # Empty set or name is always an error.
                 # TODO: To support >1 cluster per task is not implemented
-                # self.radbrpc.insertSpecificationAndTask() as called below and the radb would need to take >1
+                # self.radbrpc.insertOrUpdateSpecificationAndTask() as called below and the radb would need to take >1
                 # cluster name/ Also, there is only 1 processingClusterName in the parset, but we do not always want to
                 # pipeline process all obs outputs, or not on 1 cluster
                 logger.error(
@@ -728,7 +730,7 @@ class Specification:
 
                 # Retrieve known cluster names (not all may be a valid storage target, but we cannot know...)
                 known_cluster_set = {cluster['name'] for cluster in
-                                     self.radbrpc.getResourceGroupNames('cluster')}
+                                     self.radb.getResourceGroupNames('cluster')}
                 if cluster_name not in known_cluster_set:
                     raise Exception("skipping resource assignment for task with cluster name '" + cluster_name +
                                     "' not in known clusters " + str(known_cluster_set)) #TODO better error
@@ -973,7 +975,7 @@ class Specification:
         starttime, endtime, duration, cluster) or None
         """
         #TODO maybe this should read the spec as well? Can especially start/end times be updated outside of here?
-        task = self.radbrpc.getTask(radb_id)  # if radb_id is not None else None
+        task = self.radb.getTask(radb_id)  # if radb_id is not None else None
         if task: #TODO what is the exact structure of task see schedulerchecker 47
             self.radb_id  = task["id"] # Should be the same as radb_id, but self.radb_id might not yet be set
             self.mom_id   = task["mom_id"]
@@ -1045,7 +1047,7 @@ class Specification:
             # another service sets the parset spec in OTDB, and updated otdb task status to scheduled, which is then
             # synced back to RADB
             if self.radb_id:
-                self.radbrpc.updateTask(self.radb_id, task_status=new_status)
+                self.radb.updateTask(self.radb_id, task_status=new_status)
             # TODO what? else:
             #    self.radbrpc.updateTaskStatusForOtdbId(self.otdb_id, 'error')
         #TODO, see if we move _send_task_status_notification from resource_assigner to here?
@@ -1064,8 +1066,8 @@ class Specification:
             'insertSpecification mom_id=%s, otdb_id=%s, status=%s, task_type=%s, start_time=%s, end_time=%s '
             'cluster=%s' % (self.mom_id, self.otdb_id, self.status, self.type, self.starttime, self.endtime, self.cluster)
         )
-        result = self.radbrpc.insertSpecificationAndTask(self.mom_id, self.otdb_id, self.status, self.type, self.starttime,
-                                                         self.endtime, str(self.as_dict()), self.cluster) #TODO use internal_dict?
+        result = self.radb.insertOrUpdateSpecificationAndTask(self.mom_id, self.otdb_id, self.status, self.type, self.starttime,
+                                                              self.endtime, str(self.as_dict()), self.cluster, commit=True) #TODO use internal_dict?
 
         specification_id = result['specification_id'] # We never seem to need this again
         task_id = result['task_id']
@@ -1088,17 +1090,26 @@ class Specification:
 
         for predecessor_mom_id in predecessor_mom_ids:
             # check if the predecessor needs to be linked to this task
-            predecessor_task = self.radbrpc.getTask(mom_id=predecessor_mom_id)
+            predecessor_task = self.radb.getTask(mom_id=predecessor_mom_id)
             if predecessor_task:
                 predecessor_keys = [p.radb_id for p in self.predecessors]
                 if predecessor_task['id'] not in predecessor_keys:
                     logger.info('connecting predecessor task with mom_id=%s otdb_id=%s to its successor with mom_id=%s '
                                 'otdb_id=%s', predecessor_task['mom_id'], predecessor_task['otdb_id'], self.mom_id,
                                 self.otdb_id)
-                    spec = Specification(self.otdbrpc, self.momquery, self.radbrpc)
+                    spec = Specification(self.otdbrpc, self.momquery, self.radb)
                     spec.read_from_radb(predecessor_task['id'])
                     self.predecessors.append(spec)  # TODO this needs a try/except somewhere?
-                    self.radbrpc.insertTaskPredecessor(self.radb_id, spec.radb_id)
+                    try:
+                        self.radb.insertTaskPredecessor(self.radb_id, spec.radb_id)
+                    except PostgresDBQueryExecutionError as e:
+                        # task was already connected to predecessor. Log and continue.
+                        if 'task_predecessor_unique' in str(e):
+                            logger.info('predecessor task with mom_id=%s otdb_id=%s was already connected to its successor with mom_id=%s otdb_id=%s',
+                                        predecessor_task['mom_id'], predecessor_task['otdb_id'],
+                                        self.mom_id, self.otdb_id)
+                        else:
+                            raise
             else:
                 # Occurs when setting a pipeline to prescheduled while a predecessor has e.g. never been beyond
                 # approved, which is in principle valid. The link in the radb will be made later via processSuccessors()
@@ -1123,7 +1134,7 @@ class Specification:
 
         for successor_mom_id in successor_mom_ids:
             # check if the successor needs to be linked to this task
-            successor_task = self.radbrpc.getTask(mom_id=successor_mom_id)
+            successor_task = self.radb.getTask(mom_id=successor_mom_id)
             if successor_task:
                 if successor_task['id'] not in self.successor_ids:
                     logger.info(
@@ -1131,9 +1142,18 @@ class Specification:
                         ' otdb_id=%s', successor_task['mom_id'], successor_task['otdb_id'], self.mom_id, self.otdb_id
                     )
                     self.successor_ids.append(successor_task['id'])
-                    self.radbrpc.insertTaskPredecessor(successor_task['id'], self.radb_id)
-
-                movePipelineAfterItsPredecessors(successor_task, self.radbrpc)
+                    try:
+                        self.radb.insertTaskPredecessor(successor_task['id'], self.radb_id)
+                    except PostgresDBQueryExecutionError as e:
+                        # task was already connected to predecessor. Log and continue.
+                        if 'task_predecessor_unique' in str(e):
+                            logger.info('successor task with mom_id=%s otdb_id=%s was already connected to its predecessor with mom_id=%s otdb_id=%s',
+                                        successor_task['mom_id'], successor_task['otdb_id'],
+                                        self.mom_id, self.otdb_id)
+                        else:
+                            raise
+
+                movePipelineAfterItsPredecessors(successor_task, self.radb)
             else:
                 # Occurs when settings a obs or task to prescheduled while a successor has e.g. not yet been beyond
                 # approved, which is quite normal. The link in the radb will be made later via processPredecessors()
diff --git a/SAS/ResourceAssignment/ResourceAssigner/lib/resource_assigner.py b/SAS/ResourceAssignment/ResourceAssigner/lib/resource_assigner.py
index 45eb4b35ef9a5f0c57394cee5b5dde1b1a29db57..113d6ab2198c5930ae1c6b791c3c4b1961816bd2 100755
--- a/SAS/ResourceAssignment/ResourceAssigner/lib/resource_assigner.py
+++ b/SAS/ResourceAssignment/ResourceAssigner/lib/resource_assigner.py
@@ -32,7 +32,7 @@ from lofar.messaging.messagebus import ToBus
 from lofar.messaging import DEFAULT_BROKER, DEFAULT_BUSNAME
 from lofar.common.util import single_line_with_single_spaces
 
-from lofar.sas.resourceassignment.resourceassignmentservice.rpc import RADBRPC
+from lofar.sas.resourceassignment.database.radb import RADatabase
 from lofar.sas.otdb.otdbrpc import OTDBRPC
 
 from lofar.sas.resourceassignment.resourceassigner.config import DEFAULT_RA_NOTIFICATION_PREFIX
@@ -72,7 +72,7 @@ class ResourceAssigner(object):
         :param radb_dbcreds: the credentials to be used for accessing the RADB (default: None, which means default)
         """
 
-        self.radbrpc = RADBRPC.create(exchange=exchange, broker=broker)
+        self.radb = RADatabase(dbcreds=radb_dbcreds)
         self.rerpc = ResourceEstimatorRPC.create(exchange=exchange, broker=broker)
         self.otdbrpc = OTDBRPC.create(exchange=exchange, broker=broker)
         self.momrpc = MoMQueryRPC.create(exchange=exchange, broker=broker)
@@ -81,7 +81,7 @@ class ResourceAssigner(object):
         self.ra_notification_bus = ToBus(exchange=exchange, broker=broker)
         self.obscontrol = ObservationControlRPCClient.create(exchange=exchange, broker=broker)
 
-        self.resource_availability_checker = ResourceAvailabilityChecker(self.radbrpc)
+        self.resource_availability_checker = ResourceAvailabilityChecker(self.radb)
 
         # For the DwellScheduler instances created during run-time we store the following variables
         self.radb_creds = radb_dbcreds
@@ -97,8 +97,9 @@ class ResourceAssigner(object):
         self.close()
 
     def open(self):
-        """Open rpc connections to radb service and resource estimator service"""
-        self.radbrpc.open()
+        """Open connections to various service/buses/databases"""
+        logger.debug("resource_assigner opening all bus/db connections")
+        self.radb.connect()
         self.rerpc.open()
         self.otdbrpc.open()
         self.momrpc.open()
@@ -106,24 +107,27 @@ class ResourceAssigner(object):
         self.curpc.open()
         self.ra_notification_bus.open()
         self.obscontrol.open()
+        logger.info("resource_assigner opened all bus/db connections")
 
     def close(self):
-        """Close rpc connections to radb service and resource estimator service"""
+        """Close connections to various service/buses/databases"""
+        logger.debug("resource_assigner closing all bus/db connections")
         self.obscontrol.close()
-        self.radbrpc.close()
+        self.radb.disconnect()
         self.rerpc.close()
         self.otdbrpc.close()
         self.momrpc.close()
         self.sqrpc.close()
         self.curpc.close()
         self.ra_notification_bus.close()
+        logger.info("resource_assigner closed all bus/db connections")
 
     @property
     @cache
     def resource_types(self):
         """ Returns a dict of all the resource types, to convert name->id. """
 
-        return {rt['name']: rt['id'] for rt in self.radbrpc.getResourceTypes()}
+        return {rt['name']: rt['id'] for rt in self.radb.getResourceTypes()}
 
     def do_assignment(self, otdb_id, specification_tree):
         """
@@ -140,9 +144,9 @@ class ResourceAssigner(object):
         :raises an Exception if something unforeseen happened while scheduling
         """
 
-        logger.info(('do_assignment: otdb_id=%s specification_tree=%s' % (otdb_id, specification_tree)))
+        logger.info('do_assignment: otdb_id=%s specification_tree=%s', otdb_id, specification_tree)
 
-        spec = Specification(self.otdbrpc, self.momrpc, self.radbrpc)
+        spec = Specification(self.otdbrpc, self.momrpc, self.radb)
         spec.from_dict(specification_tree)
         spec.insert_into_radb() # TODO Move this to TaskSpecified?
 
@@ -269,7 +273,7 @@ class ResourceAssigner(object):
                                            specification_tree=specification_tree,
                                            resource_estimator=self._get_resource_estimates,
                                            resource_availability_checker=self.resource_availability_checker,
-                                           radbcreds=self.radb_creds,
+                                           radb=self.radb,
                                            min_starttime=min_starttime,
                                            max_starttime=max_starttime,
                                            duration=duration)
@@ -278,7 +282,7 @@ class ResourceAssigner(object):
                                               specification_tree=specification_tree,
                                               resource_estimator=self._get_resource_estimates,
                                               resource_availability_checker=self.resource_availability_checker,
-                                              radbcreds=self.radb_creds)
+                                              radb=self.radb)
         except Exception as e:
             logger.exception('Error in scheduler._schedule_resources: %s', e) #Why are we mentioning _schedule_resources here?
             return False
@@ -288,13 +292,14 @@ class ResourceAssigner(object):
                 (scheduler_result, changed_tasks) = scheduler.allocate_resources()
                 if not scheduler_result:
                     # try again with basic scheduler to end up with a situation with the 'normal' conflicting resources, which can then be evaluated by users
-                    with BasicScheduler(task_id=spec.radb_id,
-                                        specification_tree=specification_tree,
-                                        resource_estimator=self._get_resource_estimates,
-                                        resource_availability_checker=self.resource_availability_checker,
-                                        radbcreds=self.radb_creds) as basic_scheduler:
+                    basic_scheduler = BasicScheduler(task_id=spec.radb_id,
+                                                     specification_tree=specification_tree,
+                                                     resource_estimator=self._get_resource_estimates,
+                                                     resource_availability_checker=self.resource_availability_checker,
+                                                     radb=self.radb)
+                    with basic_scheduler:
                         (scheduler_result, changed_tasks) = basic_scheduler.allocate_resources()
-                    return scheduler_result
+                        return scheduler_result
                 elif changed_tasks:
                     for t in changed_tasks:
                         if t.status == 'aborted': #MAC_Scheduler can't handle queued right now See also schedulers.py around line 600
diff --git a/SAS/ResourceAssignment/ResourceAssigner/lib/resource_availability_checker.py b/SAS/ResourceAssignment/ResourceAssigner/lib/resource_availability_checker.py
index 08bddad5f0a01a38b65e92e10c52d17ad0ad3e55..412982ac92019c1c7967641aedd6216d311c8ce3 100644
--- a/SAS/ResourceAssignment/ResourceAssigner/lib/resource_availability_checker.py
+++ b/SAS/ResourceAssignment/ResourceAssigner/lib/resource_availability_checker.py
@@ -33,8 +33,8 @@ class CouldNotFindClaimException(Exception):
 
 
 class ResourceAvailabilityChecker(object):
-    def __init__(self, radbrpc):
-        self.radbrpc = radbrpc
+    def __init__(self, radb):
+        self.radb = radb
 
     @property
     @cache
@@ -47,7 +47,7 @@ class ResourceAvailabilityChecker(object):
               "parent_ids": list of parent resource groups
               "resource_ids": list of resources in this group. } """
 
-        memberships = self.radbrpc.getResourceGroupMemberships()
+        memberships = self.radb.getResourceGroupMemberships()
         return memberships['groups']  # resource-group-to-resource-group relations
 
     @property
@@ -55,14 +55,14 @@ class ResourceAvailabilityChecker(object):
     def resource_types(self):
         """ Returns a dict of all the resource types, to convert name->id. """
 
-        return {rt['name']: rt['id'] for rt in self.radbrpc.getResourceTypes()}
+        return {rt['name']: rt['id'] for rt in self.radb.getResourceTypes()}
 
     @property
     @cache
     def resource_claim_property_types(self):
         """ Returns a dict of all the resource claim property types, to convert name->id. """
 
-        return {rcpt['name']: rcpt['id'] for rcpt in self.radbrpc.getResourceClaimPropertyTypes()}
+        return {rcpt['name']: rcpt['id'] for rcpt in self.radb.getResourceClaimPropertyTypes()}
 
     @cache
     def _summable_property_type_ids(self):
@@ -145,8 +145,8 @@ class ResourceAvailabilityChecker(object):
         return claims
 
     def _get_current_resource_usage(self):
-        db_resource_list = self.radbrpc.getResources(include_availability=True)
-        db_resource_max_fill_ratios = self.radbrpc.getResourceAllocationConfig(sql_like_name_pattern='max_fill_ratio_%')
+        db_resource_list = self.radb.getResources(include_availability=True)
+        db_resource_max_fill_ratios = self.radb.getResourceAllocationConfig(sql_like_name_pattern='max_fill_ratio_%')
 
         self._apply_maximum_fill_ratios(db_resource_list, db_resource_max_fill_ratios)
 
diff --git a/SAS/ResourceAssignment/ResourceAssigner/lib/schedulers.py b/SAS/ResourceAssignment/ResourceAssigner/lib/schedulers.py
index 0f17dae22ddc74a22a51b7a79d8860417692480a..334008e611918f6959bcdd166e50f0da12b6cf99 100644
--- a/SAS/ResourceAssignment/ResourceAssigner/lib/schedulers.py
+++ b/SAS/ResourceAssignment/ResourceAssigner/lib/schedulers.py
@@ -47,7 +47,7 @@ class ScheduleException(Exception):
 class BasicScheduler(object):
     """ A Scheduler that allocates resources at a fixed time. Resources are searched for. """
 
-    def __init__(self, task_id, specification_tree, resource_estimator, resource_availability_checker, radbcreds=None):
+    def __init__(self, task_id, specification_tree, resource_estimator, resource_availability_checker, radb: RADatabase):
         """
         Creates a BasicScheduler instance
 
@@ -55,7 +55,7 @@ class BasicScheduler(object):
         :param specification_tree: the full specification; will be modified where needed with respect to resources
         :param resource_estimator: the ResourceEstimator function that turns a specification into resource estimates
         :param resource_availability_checker: the ResourceAvailabilityScheduler to be used by the BasicScheduler
-        :param radbcreds: ResourceAssigner database credentials. If None, the default credentials will be used
+        :param radb: a RADatabase instance.
 
         :raises AssertionError if task_id is a negative number or is None
         """
@@ -64,9 +64,7 @@ class BasicScheduler(object):
         self.specification_tree = specification_tree
         self.resource_estimator = resource_estimator
         self.resource_availability_checker = resource_availability_checker
-
-        # We need a DIRECT connection to the database in order to do client-side (that's us) transactions
-        self.radb = RADatabase(dbcreds=radbcreds)
+        self.radb = radb
 
         # Ensure a valid task_id is given, since radb.getTasks() will not raise if task_id equals None
         if task_id < 0:
@@ -92,12 +90,10 @@ class BasicScheduler(object):
         self.close()
 
     def open(self):
-        """ Open radb connection"""
-        self.radb.connect()
+        pass
 
     def close(self):
-        """ Close radb connection"""
-        self.radb.disconnect()
+        pass
 
     def allocate_resources(self):
         """
@@ -333,7 +329,7 @@ class StationScheduler(BasicScheduler):
         After scheduling, the get_stations() function returns a list of the allocated stations. """
 
     def __init__(self, task_id, specification_tree, resource_estimator, resource_availability_checker,
-                 radbcreds=None,
+                 radb: RADatabase,
                  broker=DEFAULT_BROKER):
         """
         Creates a StationScheduler instance
@@ -342,13 +338,13 @@ class StationScheduler(BasicScheduler):
         :param specification_tree: the full specification; will be modified where needed with respect to resources
         :param resource_estimator: the ResourceEstimator function that turns a specification into resource estimates
         :param resource_availability_checker: the ResourceAvailabilityChecker instance to use
-        :param radbcreds: the RADB credentials to use
+        :param radb: a RADatabase instance. 
         :param mom_busname: the MoM Query service bus name (default: 'lofar.ra.command')
         :param mom_servicename: the MoM Query service name (default: 'momqueryservice')
         :param broker: the message broker to use for send messages/RPC calls/etc.
         """
 
-        super(StationScheduler, self).__init__(task_id, specification_tree, resource_estimator, resource_availability_checker, radbcreds)
+        super(StationScheduler, self).__init__(task_id, specification_tree, resource_estimator, resource_availability_checker, radb)
 
         # For observations without a fixed station list, we need to derive one. TODO: this likely isnt the condition we want to decide on
         self.must_derive_station_list = specification_tree["task_type"] == "observation" and specification_tree["specification"]["Observation.VirtualInstrument.stationList"] == [] and specification_tree["station_requirements"]
@@ -489,7 +485,7 @@ class PriorityScheduler(StationScheduler):
         Conflict resolution is done by killing jobs with lower priority. """
 
     def __init__(self, task_id, specification_tree, resource_estimator, resource_availability_checker,
-                 radbcreds=None,
+                 radb: RADatabase,
                  broker=DEFAULT_BROKER):
         """
         Creates a PriorityScheduler instance
@@ -498,11 +494,11 @@ class PriorityScheduler(StationScheduler):
         :param specification_tree: the full specification; will be modified where needed with respect to resources
         :param resource_estimator: the ResourceEstimator function that turns a specification into resource estimates
         :param resource_availability_checker: the ResourceAvailabilityChecker instance to use
-        :param radbcreds: the RADB credentials to use
+        :param radb: a RADatabase instance. 
         :param broker: the message broker to use for send messages/RPC calls/etc.
         """
 
-        super(PriorityScheduler, self).__init__(task_id, specification_tree, resource_estimator, resource_availability_checker, radbcreds)
+        super(PriorityScheduler, self).__init__(task_id, specification_tree, resource_estimator, resource_availability_checker, radb)
 
         self.momqueryservice = MoMQueryRPC.create(broker=broker, timeout=180)
 
@@ -719,7 +715,7 @@ class DwellScheduler(PriorityScheduler):
             max_starttime,
             duration,
             resource_availability_checker,
-            radbcreds=None,
+            radb: RADatabase = None,
             broker=DEFAULT_BROKER):
         """
         Create a DwellScheduler instance
@@ -731,7 +727,7 @@ class DwellScheduler(PriorityScheduler):
         :param max_starttime: the task's desired maximum start time
         :param duration: the task's duration
         :param resource_availability_checker: the ResourceAvailabilityChecker to use
-        :param radbcreds: the RADB credentials to use
+        :param radb: a RADatabase instance. 
         :param mom_busname: the MoM Query service bus name (default: 'lofar.ra.command')
         :param mom_servicename: the MoM Query service name (default: 'momqueryservice')
         :param broker: the message broker to use for send messages/RPC calls/etc.
@@ -741,7 +737,7 @@ class DwellScheduler(PriorityScheduler):
             specification_tree=specification_tree,
             resource_estimator=resource_estimator,
             resource_availability_checker=resource_availability_checker,
-            radbcreds=radbcreds,
+            radb=radb,
             broker=broker)
 
         self.min_starttime = min_starttime
diff --git a/SAS/ResourceAssignment/ResourceAssigner/test/t_resource_availability_checker.py b/SAS/ResourceAssignment/ResourceAssigner/test/t_resource_availability_checker.py
index 1a5ba3f2542daafbcd66fb7f74e6974aa995f7fc..db2888dc11c8145df83a4b6b768ebfb3e21210e6 100755
--- a/SAS/ResourceAssignment/ResourceAssigner/test/t_resource_availability_checker.py
+++ b/SAS/ResourceAssignment/ResourceAssigner/test/t_resource_availability_checker.py
@@ -160,7 +160,7 @@ class ResourceAvailabilityCheckerTest(unittest.TestCase):
         self.addCleanup(rarpc_patcher.stop)
         self.rarpc_mock = rarpc_patcher.start()
         self.rarpc_mock.getTask.side_effect = get_task_side_effect
-        self.rarpc_mock.insertSpecificationAndTask.return_value = {
+        self.rarpc_mock.insertOrUpdateSpecificationAndTask.return_value = {
             'inserted': True,
             'specification_id': self.specification_id,
             'task_id': self.task_id
diff --git a/SAS/ResourceAssignment/ResourceAssigner/test/t_resourceassigner.py b/SAS/ResourceAssignment/ResourceAssigner/test/t_resourceassigner.py
index 014282619181b13083d2c703b72ae5dbe13e0d2a..17cfcb1e8b647371c5a851a43a87f3a1aa475ca1 100755
--- a/SAS/ResourceAssignment/ResourceAssigner/test/t_resourceassigner.py
+++ b/SAS/ResourceAssignment/ResourceAssigner/test/t_resourceassigner.py
@@ -22,41 +22,24 @@ import unittest
 from unittest import mock
 import datetime
 import sys
+from copy import deepcopy
 
 import logging
 logging.basicConfig(format='%(asctime)s %(levelname)s %(process)s %(message)s', level=logging.INFO)
 
-from lofar.sas.resourceassignment.resourceassigner.resource_assigner import ResourceAssigner, DEFAULT_RA_NOTIFICATION_PREFIX
-from lofar.sas.resourceassignment.resourceassigner.resource_availability_checker import ResourceAvailabilityChecker
+from lofar.sas.resourceassignment.resourceassigner.config import DEFAULT_RA_NOTIFICATION_PREFIX
+from lofar.sas.resourceassignment.resourceassigner.resource_assigner import ResourceAssigner
+from lofar.messaging.messagebus import TemporaryExchange
 from lofar.sas.resourceassignment.common.specification import Specification
-from lofar.parameterset import parameterset
-from lofar.common.datetimeutils import parseDatetime
 from lofar.common.util import single_line_with_single_spaces
+from lofar.messaging.messages import EventMessage
 
+from lofar.sas.resourceassignment.database.testing.radb_common_testing import RADBCommonTestMixin
 
-class TestingResourceAssigner(ResourceAssigner):
-    def __init__(self, rarpc, rerpc, otdbrpc, momrpc, curpc, sqrpc, ra_notification_bus, dwell_scheduler, obscontrol,
-                 radb_dbcreds=None):
-        # super gets not done to be able to insert mocks as early as possible otherwise the RPC block unittesting
-        self.radbrpc = rarpc
-        self.rerpc = rerpc
-        self.otdbrpc = otdbrpc
-        self.momrpc = momrpc
-        self.curpc = curpc
-        self.sqrpc = sqrpc
-        self.ra_notification_bus = ra_notification_bus
-        # Could mock ResourceAvailabilityChecker, but it is out of play already due to mocked DwellScheduler
-        self.resource_availability_checker = ResourceAvailabilityChecker(rarpc)
-        self.dwell_scheduler = dwell_scheduler
-        self.obscontrol = obscontrol
-        self.radb_creds = radb_dbcreds
-
-
-class ResourceAssignerTest(unittest.TestCase):
+class ResourceAssignerTest(RADBCommonTestMixin, unittest.TestCase):
     mom_id = 351557
     otdb_id = 1290494
     specification_id = 2323
-    state = 'prescheduled'
     task_type = 'pipeline'
 
     specification_tree = {}
@@ -78,36 +61,16 @@ class ResourceAssignerTest(unittest.TestCase):
         }
     }
 
-    approved_status = 'approved'
-    approved_otdb_id = 22
-    approved_specification_tree = {
-        'otdb_id': approved_otdb_id,
-        'task_type': 'pipeline',
-        'state': approved_status,
-        'specification': {
-            'Observation.startTime': future_start_time,
-            'Observation.stopTime': future_stop_time
-        }
-    }
-
-    cep2_specification_tree = {
-        'otdb_id': otdb_id,
-        'task_type': 'pipeline',
-        'state': 'prescheduled',
-        'specification': {
-            'Observation.startTime': future_start_time,
-            'Observation.stopTime': future_stop_time,
-            'Observation.DataProducts.Output_Pulsar.enabled': True,
-            'Observation.DataProducts.Output_Pulsar.storageClusterName': 'CEP2'
-        }
-    }
-
     mom_bug_processing_cluster_name = 'CEP2'
     mom_bug_otdb_id = 1234
     mom_bug_specification_tree = {
         'otdb_id': mom_bug_otdb_id,
+        'mom_id': None,
+        'task_id': None,
         'task_type': 'pipeline',
         'state': 'prescheduled',
+        'starttime': datetime.datetime.utcnow(),
+        'endtime': datetime.datetime.utcnow() + datetime.timedelta(minutes=1),
         'specification': {
             'Observation.startTime': future_start_time,
             'Observation.stopTime': future_stop_time,
@@ -117,32 +80,6 @@ class ResourceAssignerTest(unittest.TestCase):
         }
     }
 
-    maintenance_otdb_id = 5678
-    maintenance_specification_tree = {
-        'otdb_id': maintenance_otdb_id,
-        'task_type': 'reservation',
-        'task_subtype': 'maintenance',
-        'state': 'prescheduled',
-        'specification': {
-            'Observation.startTime': future_start_time,
-            'Observation.stopTime': future_stop_time,
-            'Observation.VirtualInstrument.stationList': ['CS001'],
-        }
-    }
-
-    projectreservation_otdb_id = 8765
-    projectreservation_specification_tree = {
-        'otdb_id': projectreservation_otdb_id,
-        'task_type': 'reservation',
-        'task_subtype': 'project',
-        'state': 'prescheduled',
-        'specification': {
-            'Observation.startTime': future_start_time,
-            'Observation.stopTime': future_stop_time,
-            'Observation.VirtualInstrument.stationList': ['CS001'],
-        }
-    }
-
     task_mom_id = 351543
     task_otdb_id = 1290472
     task_id = 2299
@@ -159,53 +96,17 @@ class ResourceAssignerTest(unittest.TestCase):
     predecessor_task_mom_id = 1
     predecessor_task_otdb_id = 2
     predecessor_task_id = 3
-    predecessor_task = {
-        "mom_id": predecessor_task_mom_id,
-        "otdb_id": predecessor_task_otdb_id,
-        "id": predecessor_task_id,
-        "endtime": datetime.datetime(2016, 3, 25, 22, 47, 31),
-        "name": "IS HBA_DUAL",
-        "predecessor_ids": [],
-        "project_mom_id": 2,
-        "project_name": "test-lofar",
-        "specification_id": 2323,
-        "starttime": datetime.datetime(2016, 3, 25, 21, 47, 31),
-        "status": "prescheduled",
-        "status_id": 350,
-        "successor_ids": [],
-        "type": "pipeline",
-        "type_id": 0
-    }
 
     successor_task_mom_id = 4
     successor_task_otdb_id = 5
     successor_task_id = 6
-    successor_task = {
-        "mom_id": successor_task_mom_id,
-        "otdb_id": successor_task_otdb_id,
-        "id": successor_task_id,
-        "endtime": datetime.datetime(2016, 3, 25, 22, 47, 31),
-        "name": "IS HBA_DUAL",
-        "predecessor_ids": [],
-        "project_mom_id": 2,
-        "project_name": "test-lofar",
-        "specification_id": 2323,
-        "starttime": datetime.datetime(2016, 3, 25, 21, 47, 31),
-        "status": "prescheduled",
-        "status_id": 350,
-        "successor_ids": [],
-        "type": "pipeline",
-        "type_id": 0
-    }
 
-    resources_with_rcus_otdb_id = 1290495
     resources_with_no_resource_types_otdb_id = 1290497
     resources_with_negative_estimates_otdb_id = 1290488
     resources_with_errors_otdb_id = 1290496
+    no_resources_otdb_id = 1290498
     resource_error1 = "error 1"
     resource_error2 = "error 2"
-    unknown_resource_type_name = "fuel"
-    unknown_resource_type_otdb_id = 123489
 
     rerpc_status = 0
     rerpc_needed_claim_for_bandwidth_size = 2
@@ -269,13 +170,9 @@ class ResourceAssignerTest(unittest.TestCase):
                 }
             }]
         },
-        str(resources_with_rcus_otdb_id): {
-                'errors': [],
-                'estimates': [{
-                        'resource_types': {'rcu': '111100010111100101101010' },
-                        'resource_count': 1,
-                        'root_resource_group': 'CS001'
-                }]
+        str(no_resources_otdb_id): {
+            'errors': [],
+            'estimates': []
         },
         str(resources_with_errors_otdb_id): {
                 'estimates': [{
@@ -309,75 +206,9 @@ class ResourceAssignerTest(unittest.TestCase):
                             }]
                 }
             }],
-        },
-        str(unknown_resource_type_otdb_id): {
-        },
-        str(maintenance_otdb_id): 
-            {'errors': [],
-                  'estimates': [{'resource_count': 1,
-                                 'resource_types': {'rcu': '111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111'},
-                                 'root_resource_group': 'CS001'},
-                                # TODO: I'm leaving this here to check the order issue in the getResourceGroupMemberships.return_value
-                                # {'resource_count': 1,
-                                #  'resource_types': {'bandwidth': 3000000000,
-                                #                     'rsp': 3904},
-                                #  'root_resource_group': 'CS001RSP0'},
-                                # {'resource_count': 1,
-                                #  'resource_types': {'bandwidth': 3000000000,
-                                #                     'rsp': 3904},
-                                #  'root_resource_group': 'CS001RSP1'},
-                                # {'resource_count': 1,
-                                #  'resource_types': {'rcu': '111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111'},
-                                #  'root_resource_group': 'CS002'},
-                                # {'resource_count': 1,
-                                #  'resource_types': {'bandwidth': 3000000000,
-                                #                     'rsp': 3904},
-                                #  'root_resource_group': 'CS002RSP0'},
-                                # {'resource_count': 1,
-                                #  'resource_types': {'bandwidth': 3000000000,
-                                #                     'rsp': 3904},
-                                #  'root_resource_group': 'CS002RSP1'},
-                                ]}
-        ,
-        str(projectreservation_otdb_id):
-            {'errors': [],
-                   'estimates': [{'resource_count': 1,
-                                 'resource_types': {'rcu': '111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111'},
-                                 'root_resource_group': 'CS001'},
-                                ]}
-    }
-
-    cep4bandwidth_resource_id = 116
-    cep4storage_resource_id = 117
-
-    storage_claim = {
-        'resource_id': cep4storage_resource_id,
-        'resource_type_id': 5,
-        'starttime': task_start_time,
-        'used_rcus': None,
-        'endtime': task_end_time + datetime.timedelta(days=365),
-        'status': 'tentative',
-        'claim_size': rerpc_needed_claim_for_storage_size,
-        'properties': [
-            {'io_type': 'output', 'type': 15, 'sap_nr': 0, 'value': 0},
-            {'io_type': 'output', 'type': 2, 'sap_nr': 0, 'value': 1},
-            {'io_type': 'output', 'type': 10, 'sap_nr': 0, 'value': 1073741824}
-        ]
-    }
-
-    bandwidth_claim = {
-        'resource_id': cep4bandwidth_resource_id,
-        'resource_type_id': 3,
-        'starttime': task_start_time,
-        'used_rcus': None,
-        'endtime': task_end_time,
-        'status': 'tentative',
-        'claim_size': rerpc_needed_claim_for_bandwidth_size,
-        'properties': []
+        }
     }
 
-    specification_claims = [bandwidth_claim, storage_claim]
-
     def reset_specification_tree(self):
         self.specification_tree = {
             'otdb_id': self.otdb_id,
@@ -393,6 +224,9 @@ class ResourceAssignerTest(unittest.TestCase):
             'duration': 60,
             'cluster': "CEP4",
             'task_subtype': 'long baseline pipeline',
+            'starttime': self.future_start_time,
+            'endtime': self.future_stop_time,
+            'storagemanager': None,
             'specification': {
                 'Observation.momID': str(self.mom_id),
                 'Observation.startTime': self.future_start_time,
@@ -431,7 +265,6 @@ class ResourceAssignerTest(unittest.TestCase):
                 'min_duration': 60,
                 'max_duration': 60,
                 'cluster': "CEP4",
-
                 'task_subtype': 'averaging pipeline',
                 'specification': {
                     'Observation.DataProducts.Output_InstrumentModel.enabled': False,
@@ -543,1079 +376,34 @@ class ResourceAssignerTest(unittest.TestCase):
         }
 
     def setUp(self):
+        super().setUp()
         self.reset_task()
 
-        def get_task_side_effect(*args, **kwargs):
-            if 'mom_id' in kwargs:
-                if kwargs['mom_id'] == self.successor_task_mom_id:
-                    return self.successor_task
-                elif kwargs['mom_id'] == self.predecessor_task_mom_id:
-                    return self.predecessor_task
-                elif kwargs['mom_id'] == self.non_existing_task_mom_id:
-                    return None
-                else:
-                    return self.task
-            else:
-                return self.task
-
-        self.successor_task_mom_ids = [self.successor_task_mom_id]
-        self.predecessor_task_mom_ids = [self.predecessor_task_mom_id]
-
-        rarpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassignmentservice.rpc.RADBRPC')
-        self.addCleanup(rarpc_patcher.stop)
-        self.rarpc_mock = rarpc_patcher.start()
-        self.rarpc_mock.getTask.side_effect = get_task_side_effect
-        self.rarpc_mock.insertSpecificationAndTask.return_value = {
-            'inserted': True,
-            'specification_id': self.specification_id,
-            'task_id': self.task_id
-        }
-        self.rarpc_mock.getResourceClaimPropertyTypes.return_value = [
-            {'id': 0, 'name': 'nr_of_is_files'},
-            {'id': 1, 'name': 'nr_of_cs_files'},
-            {'id': 2, 'name': 'nr_of_uv_files'},
-            {'id': 3, 'name': 'nr_of_im_files'},
-            {'id': 4, 'name': 'nr_of_img_files'},
-            {'id': 5, 'name': 'nr_of_pulp_files'},
-            {'id': 6, 'name': 'nr_of_cs_stokes'},
-            {'id': 7, 'name': 'nr_of_is_stokes'},
-            {'id': 8, 'name': 'is_file_size'},
-            {'id': 9, 'name': 'cs_file_size'},
-            {'id': 10, 'name': 'uv_file_size'},
-            {'id': 11, 'name': 'im_file_size'},
-            {'id': 12, 'name': 'img_file_size'},
-            {'id': 13, 'name': 'nr_of_pulp_files'},
-            {'id': 14, 'name': 'nr_of_cs_parts'},
-            {'id': 15, 'name': 'start_sb_nr'},
-            {'id': 16, 'name': 'uv_otdb_id'},
-            {'id': 17, 'name': 'cs_otdb_id'},
-            {'id': 18, 'name': 'is_otdb_id'},
-            {'id': 19, 'name': 'im_otdb_id'},
-            {'id': 20, 'name': 'img_otdb_id'},
-            {'id': 21, 'name': 'pulp_otdb_id'},
-            {'id': 22, 'name': 'is_tab_nr'},
-            {'id': 23, 'name': 'start_sbg_nr'},
-            {'id': 24, 'name': 'pulp_file_size'}
-        ]
-        self.rarpc_mock.getResourceTypes.return_value = [
-            {'id': 0, 'name': 'rsp', 'unit_id': 0, 'units': 'rsp_channel_bit'},
-            {'id': 1, 'name': 'tbb', 'unit_id': 1, 'units': 'bytes'},
-            {'id': 2, 'name': 'rcu', 'unit_id': 2, 'units': 'rcu_board'},
-            {'id': 3, 'name': 'bandwidth', 'unit_id': 3, 'units': 'bits/second'},
-            {'id': 4, 'name': 'processor', 'unit_id': 4, 'units': 'cores'},
-            {'id': 5, 'name': 'storage', 'unit_id': 1, 'units': 'bytes'},
-        ]
-        self.rarpc_mock.insertResourceClaims.return_value = {'ids': [1, 2]}
-
-        self.rarpc_mock.getResourceGroupNames.return_value = [{"name": "CEP4"}, {"name": "DRAGNET"}, {"name": "COBALT"}]
-
-        self.rarpc_mock.getResourceGroupMemberships.return_value = {'groups': [
-            {'resource_group_parent_id': None, 'resource_group_parent_name': None, 'resource_group_id': 0,
-             'resource_group_name': 'CORE', 'child_ids': [1, 2], 'parent_ids': [], 'resource_ids': [0, 1]},
-            {'resource_group_parent_id': None, 'resource_group_parent_name': None, 'resource_group_id': 3,
-             'resource_group_name': 'CS001', 'child_ids': [], 'parent_ids': [0], 'resource_ids': [212]},
-            {'resource_group_parent_id': None, 'resource_group_parent_name': None, 'resource_group_id': 1,
-             'resource_group_name': 'CEP4', 'child_ids': [], 'parent_ids': [0], 'resource_ids': [116, 117]},
-            # {'resource_group_parent_id': None, 'resource_group_parent_name': None, 'resource_group_id': 4,   # TODO: WHY DOES ORDER MATTER IN HERE???
-            #  'resource_group_name': 'CS002', 'child_ids': [], 'parent_ids': [0], 'resource_ids': [214]},     # TODO: check what happens when this is moved after e.g. CS001; also comment in CS002 in RE response
-            ],
-            'resources': [{'resource_group_parent_id': 0,
-                           'resource_group_parent_name': 'CORE',
-                           'resource_id': 0,
-                           'resource_name': 'CS001',
-                           'parent_group_ids': []},
-                          {'resource_group_parent_id': 0,
-                           'resource_group_parent_name': 'CORE',
-                           'resource_id': 1,
-                           'resource_name': 'CS002',
-                           'parent_group_ids': []},
-                          {'resource_group_parent_id': 1,
-                           'resource_group_parent_name': 'CEP4',
-                           'resource_id': 2,
-                           'resource_name': 'CEP4_storage:/data',
-                           'parent_group_ids': []}]}
-
-        # incomplete response but good enough for tests
-        self.rarpc_mock.getResources.return_value = [
-            {'id': 0, 'name': 'cpunode01_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 1, 'name': 'cpunode01_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 2, 'name': 'cpunode02_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 3, 'name': 'cpunode02_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 4, 'name': 'cpunode03_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 5, 'name': 'cpunode03_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 6, 'name': 'cpunode04_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 7, 'name': 'cpunode04_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 8, 'name': 'cpunode05_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 9, 'name': 'cpunode05_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 10, 'name': 'cpunode06_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 11, 'name': 'cpunode06_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 12, 'name': 'cpunode07_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 13, 'name': 'cpunode07_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 14, 'name': 'cpunode08_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 15, 'name': 'cpunode08_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 16, 'name': 'cpunode09_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 17, 'name': 'cpunode09_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 18, 'name': 'cpunode10_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 19, 'name': 'cpunode10_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 20, 'name': 'cpunode11_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 21, 'name': 'cpunode11_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 22, 'name': 'cpunode12_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 23, 'name': 'cpunode12_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 24, 'name': 'cpunode13_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 25, 'name': 'cpunode13_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 26, 'name': 'cpunode14_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 27, 'name': 'cpunode14_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 28, 'name': 'cpunode15_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 29, 'name': 'cpunode15_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 30, 'name': 'cpunode16_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 31, 'name': 'cpunode16_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 32, 'name': 'cpunode17_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 33, 'name': 'cpunode17_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 34, 'name': 'cpunode18_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 35, 'name': 'cpunode18_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 36, 'name': 'cpunode19_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 37, 'name': 'cpunode19_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 38, 'name': 'cpunode20_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 39, 'name': 'cpunode20_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 40, 'name': 'cpunode21_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 41, 'name': 'cpunode21_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 42, 'name': 'cpunode22_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 43, 'name': 'cpunode22_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 44, 'name': 'cpunode23_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 45, 'name': 'cpunode23_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 46, 'name': 'cpunode24_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 47, 'name': 'cpunode24_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 48, 'name': 'cpunode25_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 49, 'name': 'cpunode25_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 50, 'name': 'cpunode26_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 51, 'name': 'cpunode26_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 52, 'name': 'cpunode27_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 53, 'name': 'cpunode27_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 54, 'name': 'cpunode28_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 55, 'name': 'cpunode28_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 56, 'name': 'cpunode29_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 57, 'name': 'cpunode29_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 58, 'name': 'cpunode30_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 59, 'name': 'cpunode30_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 60, 'name': 'cpunode31_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 61, 'name': 'cpunode31_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 62, 'name': 'cpunode32_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 63, 'name': 'cpunode32_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 64, 'name': 'cpunode33_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 65, 'name': 'cpunode33_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 66, 'name': 'cpunode34_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 67, 'name': 'cpunode34_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 68, 'name': 'cpunode35_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 69, 'name': 'cpunode35_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 70, 'name': 'cpunode36_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 71, 'name': 'cpunode36_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 72, 'name': 'cpunode37_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 73, 'name': 'cpunode37_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 74, 'name': 'cpunode38_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 75, 'name': 'cpunode38_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 76, 'name': 'cpunode39_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 77, 'name': 'cpunode39_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 78, 'name': 'cpunode40_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 79, 'name': 'cpunode40_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 80, 'name': 'cpunode41_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 81, 'name': 'cpunode41_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 82, 'name': 'cpunode42_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 83, 'name': 'cpunode42_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 84, 'name': 'cpunode43_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 85, 'name': 'cpunode43_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 86, 'name': 'cpunode44_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 87, 'name': 'cpunode44_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 88, 'name': 'cpunode45_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 89, 'name': 'cpunode45_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 90, 'name': 'cpunode46_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 91, 'name': 'cpunode46_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 92, 'name': 'cpunode47_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 93, 'name': 'cpunode47_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 94, 'name': 'cpunode48_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 95, 'name': 'cpunode48_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 96, 'name': 'cpunode49_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 97, 'name': 'cpunode49_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 98, 'name': 'cpunode50_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 99, 'name': 'cpunode50_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 100, 'name': 'cbt001_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 101, 'name': 'cbt001_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 102, 'name': 'cbt002_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 103, 'name': 'cbt002_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 104, 'name': 'cbt003_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 105, 'name': 'cbt003_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 106, 'name': 'cbt004_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 107, 'name': 'cbt004_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 108, 'name': 'cbt005_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 109, 'name': 'cbt005_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 110, 'name': 'cbt006_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 111, 'name': 'cbt006_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 112, 'name': 'cbt007_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 113, 'name': 'cbt007_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 114, 'name': 'cbt008_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 115, 'name': 'cbt008_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4,
-             'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 116, 'name': 'CEP4_bandwidth:/data', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 117, 'name': 'CEP4_storage:/data', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 118, 'name': 'dragproc_bandwidth:/data', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 119, 'name': 'dragproc_storage:/data', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 120, 'name': 'drg01_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 121, 'name': 'drg01_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 122, 'name': 'drg01_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 123, 'name': 'drg01_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 124, 'name': 'drg02_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 125, 'name': 'drg02_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 126, 'name': 'drg02_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 127, 'name': 'drg02_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 128, 'name': 'drg03_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 129, 'name': 'drg03_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 130, 'name': 'drg03_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 131, 'name': 'drg03_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 132, 'name': 'drg04_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 133, 'name': 'drg04_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 134, 'name': 'drg04_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 135, 'name': 'drg04_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 136, 'name': 'drg05_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 137, 'name': 'drg05_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 138, 'name': 'drg05_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 139, 'name': 'drg05_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 140, 'name': 'drg06_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 141, 'name': 'drg06_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 142, 'name': 'drg06_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 143, 'name': 'drg06_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 144, 'name': 'drg07_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 145, 'name': 'drg07_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 146, 'name': 'drg07_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 147, 'name': 'drg07_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 148, 'name': 'drg08_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 149, 'name': 'drg08_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 150, 'name': 'drg08_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 151, 'name': 'drg08_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 152, 'name': 'drg09_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 153, 'name': 'drg09_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 154, 'name': 'drg09_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 155, 'name': 'drg09_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 156, 'name': 'drg10_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 157, 'name': 'drg10_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 158, 'name': 'drg10_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 159, 'name': 'drg10_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 160, 'name': 'drg11_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 161, 'name': 'drg11_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 162, 'name': 'drg11_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 163, 'name': 'drg11_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 164, 'name': 'drg12_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 165, 'name': 'drg12_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 166, 'name': 'drg12_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 167, 'name': 'drg12_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 168, 'name': 'drg13_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 169, 'name': 'drg13_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 170, 'name': 'drg13_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 171, 'name': 'drg13_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 172, 'name': 'drg14_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 173, 'name': 'drg14_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 174, 'name': 'drg14_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 175, 'name': 'drg14_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 176, 'name': 'drg15_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 177, 'name': 'drg15_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 178, 'name': 'drg15_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 179, 'name': 'drg15_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 180, 'name': 'drg16_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 181, 'name': 'drg16_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 182, 'name': 'drg16_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 183, 'name': 'drg16_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 184, 'name': 'drg17_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 185, 'name': 'drg17_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 186, 'name': 'drg17_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 187, 'name': 'drg17_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 188, 'name': 'drg18_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 189, 'name': 'drg18_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 190, 'name': 'drg18_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 191, 'name': 'drg18_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 192, 'name': 'drg19_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 193, 'name': 'drg19_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 194, 'name': 'drg19_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 195, 'name': 'drg19_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 196, 'name': 'drg20_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 197, 'name': 'drg20_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 198, 'name': 'drg20_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 199, 'name': 'drg20_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 200, 'name': 'drg21_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 201, 'name': 'drg21_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 202, 'name': 'drg21_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 203, 'name': 'drg21_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 204, 'name': 'drg22_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 205, 'name': 'drg22_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 206, 'name': 'drg22_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 207, 'name': 'drg22_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 208, 'name': 'drg23_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 209, 'name': 'drg23_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3,
-             'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 210, 'name': 'drg23_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 211, 'name': 'drg23_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1,
-             'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 212, 'name': 'CS001rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 213, 'name': 'CS001tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 214, 'name': 'CS002rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 215, 'name': 'CS002tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 216, 'name': 'CS003rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 217, 'name': 'CS003tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 218, 'name': 'CS004rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 219, 'name': 'CS004tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 220, 'name': 'CS005rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 221, 'name': 'CS005tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 222, 'name': 'CS006rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 223, 'name': 'CS006tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 224, 'name': 'CS007rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 225, 'name': 'CS007tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 226, 'name': 'CS011rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 227, 'name': 'CS011tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 228, 'name': 'CS013rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 229, 'name': 'CS013tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 230, 'name': 'CS017rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 231, 'name': 'CS017tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 232, 'name': 'CS021rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 233, 'name': 'CS021tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 234, 'name': 'CS024rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 235, 'name': 'CS024tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 236, 'name': 'CS026rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 237, 'name': 'CS026tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 238, 'name': 'CS028rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 239, 'name': 'CS028tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 240, 'name': 'CS030rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 241, 'name': 'CS030tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 242, 'name': 'CS031rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 243, 'name': 'CS031tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 244, 'name': 'CS032rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 245, 'name': 'CS032tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 246, 'name': 'CS101rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 247, 'name': 'CS101tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 248, 'name': 'CS103rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 249, 'name': 'CS103tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 250, 'name': 'CS201rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 251, 'name': 'CS201tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 252, 'name': 'CS301rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 253, 'name': 'CS301tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 254, 'name': 'CS302rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 255, 'name': 'CS302tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 256, 'name': 'CS401rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 257, 'name': 'CS401tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 258, 'name': 'CS501rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 259, 'name': 'CS501tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 260, 'name': 'RS106rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 261, 'name': 'RS106tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 262, 'name': 'RS205rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 263, 'name': 'RS205tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 264, 'name': 'RS208rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 265, 'name': 'RS208tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 266, 'name': 'RS210rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 267, 'name': 'RS210tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 268, 'name': 'RS305rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 269, 'name': 'RS305tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 270, 'name': 'RS306rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 271, 'name': 'RS306tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 272, 'name': 'RS307rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 273, 'name': 'RS307tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 274, 'name': 'RS310rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 275, 'name': 'RS310tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 276, 'name': 'RS406rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 277, 'name': 'RS406tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 278, 'name': 'RS407rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 279, 'name': 'RS407tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 280, 'name': 'RS408rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 281, 'name': 'RS408tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 282, 'name': 'RS409rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 283, 'name': 'RS409tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 284, 'name': 'RS503rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 285, 'name': 'RS503tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 286, 'name': 'RS508rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 287, 'name': 'RS508tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 288, 'name': 'RS509rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 289, 'name': 'RS509tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 290, 'name': 'DE601rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 291, 'name': 'DE601tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 292, 'name': 'DE602rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 293, 'name': 'DE602tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 294, 'name': 'DE603rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 295, 'name': 'DE603tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 296, 'name': 'DE604rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 297, 'name': 'DE604tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 298, 'name': 'DE605rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 299, 'name': 'DE605tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 300, 'name': 'FR606rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 301, 'name': 'FR606tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 302, 'name': 'SE607rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 303, 'name': 'SE607tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 304, 'name': 'UK608rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 305, 'name': 'UK608tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 306, 'name': 'DE609rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 307, 'name': 'DE609tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 308, 'name': 'PL610rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 309, 'name': 'PL610tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 310, 'name': 'PL611rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 311, 'name': 'PL611tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 312, 'name': 'PL612rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 313, 'name': 'PL612tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 314, 'name': 'IE613rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 315, 'name': 'IE613tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 316, 'name': 'IS614rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 317, 'name': 'IS614tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 318, 'name': 'TEST1rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 319, 'name': 'TEST1tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 320, 'name': 'CS001chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 321, 'name': 'CS001bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 322, 'name': 'CS001chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 323, 'name': 'CS001bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 324, 'name': 'CS002chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 325, 'name': 'CS002bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 326, 'name': 'CS002chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 327, 'name': 'CS002bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 328, 'name': 'CS003chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 329, 'name': 'CS003bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 330, 'name': 'CS003chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 331, 'name': 'CS003bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 332, 'name': 'CS004chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 333, 'name': 'CS004bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 334, 'name': 'CS004chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 335, 'name': 'CS004bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 336, 'name': 'CS005chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 337, 'name': 'CS005bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 338, 'name': 'CS005chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 339, 'name': 'CS005bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 340, 'name': 'CS006chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 341, 'name': 'CS006bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 342, 'name': 'CS006chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 343, 'name': 'CS006bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 344, 'name': 'CS007chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 345, 'name': 'CS007bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 346, 'name': 'CS007chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 347, 'name': 'CS007bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 348, 'name': 'CS011chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 349, 'name': 'CS011bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 350, 'name': 'CS011chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 351, 'name': 'CS011bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 352, 'name': 'CS013chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 353, 'name': 'CS013bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 354, 'name': 'CS013chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 355, 'name': 'CS013bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 356, 'name': 'CS017chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 357, 'name': 'CS017bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 358, 'name': 'CS017chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 359, 'name': 'CS017bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 360, 'name': 'CS021chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 361, 'name': 'CS021bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 362, 'name': 'CS021chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 363, 'name': 'CS021bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 364, 'name': 'CS024chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 365, 'name': 'CS024bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 366, 'name': 'CS024chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 367, 'name': 'CS024bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 368, 'name': 'CS026chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 369, 'name': 'CS026bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 370, 'name': 'CS026chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 371, 'name': 'CS026bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 372, 'name': 'CS028chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 373, 'name': 'CS028bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 374, 'name': 'CS028chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 375, 'name': 'CS028bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 376, 'name': 'CS030chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 377, 'name': 'CS030bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 378, 'name': 'CS030chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 379, 'name': 'CS030bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 380, 'name': 'CS031chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 381, 'name': 'CS031bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 382, 'name': 'CS031chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 383, 'name': 'CS031bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 384, 'name': 'CS032chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 385, 'name': 'CS032bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 386, 'name': 'CS032chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 387, 'name': 'CS032bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 388, 'name': 'CS101chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 389, 'name': 'CS101bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 390, 'name': 'CS101chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 391, 'name': 'CS101bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 392, 'name': 'CS103chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 393, 'name': 'CS103bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 394, 'name': 'CS103chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 395, 'name': 'CS103bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 396, 'name': 'CS201chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 397, 'name': 'CS201bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 398, 'name': 'CS201chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 399, 'name': 'CS201bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 400, 'name': 'CS301chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 401, 'name': 'CS301bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 402, 'name': 'CS301chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 403, 'name': 'CS301bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 404, 'name': 'CS302chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 405, 'name': 'CS302bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 406, 'name': 'CS302chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 407, 'name': 'CS302bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 408, 'name': 'CS401chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 409, 'name': 'CS401bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 410, 'name': 'CS401chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 411, 'name': 'CS401bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 412, 'name': 'CS501chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 413, 'name': 'CS501bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 414, 'name': 'CS501chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 415, 'name': 'CS501bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 416, 'name': 'RS106chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 417, 'name': 'RS106bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 418, 'name': 'RS205chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 419, 'name': 'RS205bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 420, 'name': 'RS208chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 421, 'name': 'RS208bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 422, 'name': 'RS210chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 423, 'name': 'RS210bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 424, 'name': 'RS305chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 425, 'name': 'RS305bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 426, 'name': 'RS306chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 427, 'name': 'RS306bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 428, 'name': 'RS307chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 429, 'name': 'RS307bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 430, 'name': 'RS310chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 431, 'name': 'RS310bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 432, 'name': 'RS406chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 433, 'name': 'RS406bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 434, 'name': 'RS407chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 435, 'name': 'RS407bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 436, 'name': 'RS408chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 437, 'name': 'RS408bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 438, 'name': 'RS409chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 439, 'name': 'RS409bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 440, 'name': 'RS503chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 441, 'name': 'RS503bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 442, 'name': 'RS508chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 443, 'name': 'RS508bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 444, 'name': 'RS509chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 445, 'name': 'RS509bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 446, 'name': 'DE601chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 447, 'name': 'DE601bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 448, 'name': 'DE602chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 449, 'name': 'DE602bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 450, 'name': 'DE603chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 451, 'name': 'DE603bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 452, 'name': 'DE604chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 453, 'name': 'DE604bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 454, 'name': 'DE605chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 455, 'name': 'DE605bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 456, 'name': 'FR606chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 457, 'name': 'FR606bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 458, 'name': 'SE607chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 459, 'name': 'SE607bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 460, 'name': 'UK608chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 461, 'name': 'UK608bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 462, 'name': 'DE609chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 463, 'name': 'DE609bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 464, 'name': 'PL610chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 465, 'name': 'PL610bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 466, 'name': 'PL611chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 467, 'name': 'PL611bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 468, 'name': 'PL612chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 469, 'name': 'PL612bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 470, 'name': 'IE613chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 471, 'name': 'IE613bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 472, 'name': 'IS614chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 473, 'name': 'IS614bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 474, 'name': 'TEST1chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1},
-            {'id': 475, 'name': 'TEST1bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second',
-             'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}
-        ]
-
-        self.rarpc_mock.getResourceClaims.return_value = []
-
-        self.rarpc_mock.getResourceAllocationConfig.return_value = [
-            {'name': 'max_fill_ratio_CEP4_storage', 'value': 0.85}, {'name': 'claim_timeout', 'value': 172800},
-            {'name': 'min_inter_task_delay', 'value': 60}, {'name': 'max_fill_ratio_CEP4_bandwidth', 'value': 0.75}
-        ]
+        self.tmp_exchange = TemporaryExchange(__package__)
+        self.tmp_exchange.open()
+        self.addCleanup(self.tmp_exchange.close)
 
         def rerpc_mock_get_estimated_resources(specification_tree):
             otdb_id = specification_tree['otdb_id']
             return self.rerpc_replymessage[str(otdb_id)]
 
-        rerpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassignmentestimator.rpc.ResourceEstimatorRPC')
+        rerpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.ResourceEstimatorRPC')
         self.addCleanup(rerpc_patcher.stop)
         self.rerpc_mock = rerpc_patcher.start()
+        self.rerpc_mock.create.side_effect = lambda **kwargs: self.rerpc_mock # make factory method 'create' return the mock instance
         self.rerpc_mock.get_estimated_resources.side_effect = rerpc_mock_get_estimated_resources
 
-        otdbrpc_patcher = mock.patch('lofar.sas.otdb.otdbrpc')
+        otdbrpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.OTDBRPC')
         self.addCleanup(otdbrpc_patcher.stop)
         self.otdbrpc_mock = otdbrpc_patcher.start()
+        self.otdbrpc_mock.create.side_effect = lambda **kwargs: self.otdbrpc_mock # make factory method 'create' return the mock instance
 
-        momrpc_patcher = mock.patch('lofar.mom.momqueryservice.momqueryrpc')
+        momrpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.MoMQueryRPC')
         self.addCleanup(momrpc_patcher.stop)
         self.momrpc_mock = momrpc_patcher.start()
-        self.momrpc_mock.getPredecessorIds.return_value = {str(self.task_mom_id): self.predecessor_task_mom_ids}
-        self.momrpc_mock.getSuccessorIds.return_value = {str(self.task_mom_id): self.successor_task_mom_ids}
+        self.momrpc_mock.create.side_effect = lambda **kwargs: self.momrpc_mock # make factory method 'create' return the mock instance
+        self.momrpc_mock.getPredecessorIds.return_value = {str(self.task_mom_id): [self.predecessor_task_mom_id]}
+        self.momrpc_mock.getSuccessorIds.return_value = {str(self.task_mom_id): [self.successor_task_mom_id]}
         self.momrpc_mock.get_time_restrictions.return_value = {
             "minStartTime": self.task_minstarttime.strftime('%Y-%m-%dT%H:%M:%S'),
             "maxEndTime": self.task_maxendtime.strftime('%Y-%m-%dT%H:%M:%S'),
@@ -1623,96 +411,56 @@ class ResourceAssignerTest(unittest.TestCase):
             "maxDuration": str(self.task_maxduration)
         }
 
-        curpc_patcher = mock.patch('lofar.sas.datamanagement.cleanup.rpc')
+        curpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.CleanupRPC')
         self.addCleanup(curpc_patcher.stop)
         self.curpc_mock = curpc_patcher.start()
+        self.curpc_mock.create.side_effect = lambda **kwargs: self.curpc_mock # make factory method 'create' return the mock instance
+        self.curpc_mock.removeTaskData.return_value = {'deleted': True, 'message': ""}
 
-        sqrpc_patcher = mock.patch('lofar.sas.datamanagement.storagequery.rpc')
+        sqrpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.StorageQueryRPC')
         self.addCleanup(sqrpc_patcher.stop)
         self.sqrpc_mock = sqrpc_patcher.start()
+        self.sqrpc_mock.create.side_effect = lambda **kwargs: self.sqrpc_mock # make factory method 'create' return the mock instance
+        self.sqrpc_mock.getDiskUsageForOTDBId.return_value = {'found': True, 'disk_usage': 10}
 
-        ra_notification_bus_patcher = mock.patch('lofar.messaging.messagebus')
-        self.addCleanup(ra_notification_bus_patcher.stop)
-        self.ra_notification_bus_mock = ra_notification_bus_patcher.start()
-
-        logger_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.logger')
-        self.addCleanup(logger_patcher.stop)
-        self.logger_mock = logger_patcher.start()
-
-        dwell_scheduler_patcher = mock.patch(
-            'lofar.sas.resourceassignment.resourceassigner.resource_assigner.DwellScheduler'
-        )
-        self.addCleanup(dwell_scheduler_patcher.stop)
-        self.dwell_scheduler_mock = dwell_scheduler_patcher.start()
-        self.dwell_scheduler_mock().allocate_resources.return_value = (True, [])
-
-        prio_scheduler_patcher = mock.patch(
-            'lofar.sas.resourceassignment.resourceassigner.resource_assigner.PriorityScheduler'
-        )
-        self.addCleanup(prio_scheduler_patcher.stop)
-        self.prio_scheduler_mock = prio_scheduler_patcher.start()
-        self.prio_scheduler_mock().allocate_resources.return_value = (True, [])
-
-        obscontrol_patcher = mock.patch(
-            'lofar.sas.resourceassignment.resourceassigner.resource_assigner.ObservationControlRPCClient.abort_observation'
-        )
+        obscontrol_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.ObservationControlRPCClient')
         self.addCleanup(obscontrol_patcher.stop)
         self.obscontrol_mock = obscontrol_patcher.start()
+        self.obscontrol_mock.create.side_effect = lambda **kwargs: self.obscontrol_mock # make factory method 'create' return the mock instance
 
-        self.spec_mock = mock.MagicMock()
-        self.spec_mock.status = 'prescheduled'
-        self.spec_mock.radb_id = self.task_id
-        self.spec_mock.mom_id = self.task_mom_id
-        self.spec_mock.otdb_id = self.task_otdb_id
-        self.spec_mock.type = self.task_type
-        self.spec_mock.duration = datetime.timedelta(seconds=60)
-        self.spec_mock.starttime = datetime.datetime.utcnow()
-        self.spec_mock.endtime = datetime.datetime.utcnow()
-        self.spec_mock.min_starttime = datetime.datetime.utcnow()
-        self.spec_mock.max_endtime = datetime.datetime.utcnow()
-        self.spec_mock.calculate_dwell_values.return_value = (self.spec_mock.min_starttime,
-                                                              self.spec_mock.max_endtime,
-                                                              self.spec_mock.duration)
-
-        specification_patcher = mock.patch(
-            'lofar.sas.resourceassignment.resourceassigner.resource_assigner.Specification')
-        self.addCleanup(specification_patcher.stop)
-        self.specification_mock = specification_patcher.start()
-        self.specification_mock.return_value = self.spec_mock
+        ra_notification_bus_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.ToBus')
+        self.addCleanup(ra_notification_bus_patcher.stop)
+        self.ra_notification_bus_mock = ra_notification_bus_patcher.start()
 
         # Select logger output to see
         def myprint(s, *args):
             print(s % args if args else s, file=sys.stderr)
 
-        #self.logger_mock.debug.side_effect = myprint
+        logger_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.logger')
+        self.addCleanup(logger_patcher.stop)
+        self.logger_mock = logger_patcher.start()
         self.logger_mock.info.side_effect = myprint
         self.logger_mock.warn.side_effect = myprint
         self.logger_mock.error.side_effect = myprint
 
-        self.resource_assigner = TestingResourceAssigner(self.rarpc_mock, self.rerpc_mock,
-                                                         self.otdbrpc_mock, self.momrpc_mock,
-                                                         self.curpc_mock, self.sqrpc_mock,
-                                                         self.ra_notification_bus_mock,
-                                                         self.dwell_scheduler_mock,
-                                                         self.obscontrol_mock)
+        self.resource_assigner = ResourceAssigner(exchange=self.tmp_exchange.address, radb_dbcreds=self.dbcreds)
+        self.addCleanup(self.resource_assigner.close)
 
         self.reset_specification_tree()
 
     def assert_all_services_opened(self):
-        self.assertTrue(self.rarpc_mock.open.called, "RARPC.open was not called")
-        self.assertTrue(self.rerpc_mock.open.called, "RPC.open was not called")
+        self.assertTrue(self.rerpc_mock.open.called, "ResourceEstimatorRPC.open was not called")
         self.assertTrue(self.otdbrpc_mock.open.called, "OTDBRPC.open was not called")
         self.assertTrue(self.momrpc_mock.open.called, "MOMRPC.open was not called")
         self.assertTrue(self.curpc_mock.open.called, "CURPC.open was not called")
-        self.assertTrue(self.ra_notification_bus_mock.open.called, "ra_notification_bus.open was not called")
+        self.assertTrue(any(mc for mc in self.ra_notification_bus_mock.mock_calls if 'open()' in str(mc)), "ra_notification_bus.open was not called")
 
     def assert_all_services_closed(self):
-        self.assertTrue(self.rarpc_mock.close.called, "RARPC.close was not called")
-        self.assertTrue(self.rerpc_mock.close.called, "RPC.close was not called")
+        self.assertTrue(self.rerpc_mock.close.called, "ResourceEstimatorRPC.close was not called")
         self.assertTrue(self.otdbrpc_mock.close.called, "OTDBRPC.close was not called")
         self.assertTrue(self.momrpc_mock.close.called, "MOMRPC.close was not called")
         self.assertTrue(self.curpc_mock.close.called, "CURPC.close was not called")
-        self.assertTrue(self.ra_notification_bus_mock.close.called, "ra_notification_bus.close was not called")
+        self.assertTrue(any(mc for mc in self.ra_notification_bus_mock.mock_calls if 'close()' in str(mc)), "ra_notification_bus.close was not called")
 
     def test_open_opens_all_services(self):
         self.resource_assigner.open()
@@ -1725,9 +473,7 @@ class ResourceAssignerTest(unittest.TestCase):
         self.assert_all_services_closed()
 
     def test_contextManager_opens_and_closes_all_services(self):
-        with TestingResourceAssigner(self.rarpc_mock, self.rerpc_mock, self.otdbrpc_mock, self.momrpc_mock,
-                                     self.curpc_mock, self.sqrpc_mock, self.ra_notification_bus_mock,
-                                     self.dwell_scheduler_mock, self.obscontrol_mock):
+        with self.resource_assigner:
             self.assert_all_services_opened()
 
         self.assert_all_services_closed()
@@ -1735,10 +481,8 @@ class ResourceAssignerTest(unittest.TestCase):
     def test_do_assignment_logs_specification(self):
         self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
 
-        self.logger_mock.info.assert_any_call('do_assignment: otdb_id=%s specification_tree=%s' % (
-            self.specification_tree['otdb_id'],
-            self.specification_tree
-        ))
+        self.logger_mock.info.assert_any_call('do_assignment: otdb_id=%s specification_tree=%s',
+                                              self.specification_tree['otdb_id'], self.specification_tree)
 
     def test_do_assignment_log_non_approved_or_prescheduled_states(self):
         otdb_id = self.non_approved_or_prescheduled_otdb_id
@@ -1754,22 +498,18 @@ class ResourceAssignerTest(unittest.TestCase):
                 (otdb_id, status, assignable_task_states_str))
 
     def test_do_assignment_approved_task_should_not_be_rescheduled(self):
-        self.spec_mock.status = 'approved'
-
-        self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
-
-        self.logger_mock.info.assert_any_call('Task otdb_id=%s is only approved, no resource assignment needed yet' %
-                                              self.specification_tree['otdb_id'])
-
-    def test_do_assignment_inserts_specification_and_task_in_radb(self):
-        self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
+        otdb_id = self.specification_tree['otdb_id']
+        # assure task is not known yet
+        self.assertIsNone(self.radb.getTask(otdb_id=otdb_id))
 
-        self.spec_mock.insert_into_radb.assert_called()
+        self.specification_tree['status'] = 'approved'
+        self.resource_assigner.do_assignment(otdb_id, self.specification_tree)
 
-    def test_do_assignment_fills_specification_based_on_specication_dict(self):
-        self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
+        # assure task is known now, and scheduled
+        self.assertIsNotNone(self.radb.getTask(otdb_id=otdb_id))
+        self.assertEqual('approved', self.radb.getTask(otdb_id=otdb_id)['status'])
 
-        self.spec_mock.from_dict.assert_called_with(self.specification_tree)
+        self.logger_mock.info.assert_any_call('Task otdb_id=%s is only approved, no resource assignment needed yet' % otdb_id)
 
     def freeze_time_one_day_in_the_future(self, datetime_mock):
         now = datetime.datetime.utcnow() + datetime.timedelta(days=1)
@@ -1787,52 +527,35 @@ class ResourceAssignerTest(unittest.TestCase):
 
         self.rerpc_mock.get_estimated_resources.any_calls_with(self.specification_tree)
 
-    # def test_do_assignment_logs_when_otdb_id_not_needed_resources(self):
-    #     self.spec_mock.radb_id = self.otdb_id + 11
-    #     self.specification_tree["otdb_id"] = self.otdb_id + 11
-    #
-    #     self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
-    #
-    #     self.logger_mock.error.assert_any_call(
-    #         "An exception occurred while obtaining resource estimates. Exception=no otdb_id %s found in estimator results %s" %
-    #         (self.otdb_id + 11, self.rerpc_replymessage)
-    #     )
-
     def test_do_assignment_puts_spec_to_error_when_resource_estimation_gives_an_error(self):
-        self.dwell_scheduler_mock().allocate_resources.return_value = False
-        self.specification_tree["otdb_id"] = self.otdb_id + 11
+        with mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.PriorityScheduler') as scheduler_mock:
+            scheduler_mock.allocate_resources.return_value = False
 
-        self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
+            self.specification_tree["otdb_id"] = self.otdb_id + 11
+            self.specification_tree['status'] = 'prescheduled'
+            self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
 
-        self.spec_mock.set_status.assert_called_with('error')
+            task = self.radb.getTask(otdb_id=self.specification_tree["otdb_id"])
+            self.assertEqual('error', task['status'])
 
     def test_do_assignment_should_not_claim_resouces_when_otdb_id_not_needed_resources(self):
-        self.specification_tree["otdb_id"] = self.otdb_id + 1
+        self.specification_tree["otdb_id"] = self.no_resources_otdb_id
 
+        self.specification_tree['status'] = 'prescheduled'
         self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
 
-        self.rarpc_mock.insertResourceClaims.assert_not_called()
-
-    # def test_do_assignment_logs_when_task_type_not_in_needed_resources(self):
-    #     wrong_task_type = "observation"
-    #     self.spec_mock.type = wrong_task_type
-    #     self.specification_tree["task_type"] = wrong_task_type
-    #
-    #     self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
-    #
-    #     self.logger_mock.error.assert_any_call(
-    #         "An exception occurred while obtaining resource estimates. Exception=no task type %s found in estimator results %s" %
-    #         (wrong_task_type, self.rerpc_replymessage[str(self.otdb_id)])
-    #     )
+        task = self.radb.getTask(otdb_id=self.specification_tree["otdb_id"])
+        self.assertEqual([], self.radb.getResourceClaims(task_ids=task['id']))
 
     def test_do_assignment_should_not_claim_resources_when_task_type_not_in_needed_resources(self):
         wrong_task_type = "observation"
-        self.spec_mock.type = wrong_task_type
         self.specification_tree["task_type"] = wrong_task_type
-
+        self.specification_tree['status'] = 'prescheduled'
         self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
 
-        self.rarpc_mock.insertResourceClaims.assert_not_called()
+        task = self.radb.getTask(otdb_id=self.specification_tree["otdb_id"])
+        self.assertEqual('error', task['status'])
+        self.assertEqual([], self.radb.getResourceClaims(task_ids=task['id']))
 
     def test_do_assignment_should_log_single_errors_in_needed_resources(self):
         self.specification_tree["otdb_id"] = self.resources_with_errors_otdb_id
@@ -1859,76 +582,122 @@ class ResourceAssignerTest(unittest.TestCase):
             self.resource_assigner._get_resource_estimates(self.specification_tree)
 
     def ra_notification_bus_send_called_with(self, content, subject):
-        for call in self.ra_notification_bus_mock.send.call_args_list:
-            if call[0][0].subject == subject and call[0][0].content == content:
-                return True
+        for call in self.ra_notification_bus_mock().send.call_args_list:
+            if isinstance(call[0][0], EventMessage):
+                msg = call[0][0]
+                if msg.subject == subject and msg.content == content:
+                    return True
         return False
 
     def test_do_assignment_notifies_bus_when_it_was_unable_to_schedule_Conflict(self):
-        content = {'radb_id': self.task_id, 'otdb_id': self.task_otdb_id, 'mom_id': self.task_mom_id}
-        subject = 'Task' + 'Conflict'
-
-        self.spec_mock.status = 'conflict'
-        self.dwell_scheduler_mock().allocate_resources.return_value = False
-
-        self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
-
-        self.assertBusNotificationAndLogging(content, subject)
-
-    def test_do_assignment_should_read_spec_from_radb(self):
-        self.spec_mock.status = 'conflict'
-        self.dwell_scheduler_mock().allocate_resources.return_value = False
+        # prepare: insert a blocking task with a huge claim on storage (directly via the radb, not via the resource_assigner)
+        task_id = self.radb.insertOrUpdateSpecificationAndTask(9876, 9876, 'prescheduled', 'observation',
+                                                       datetime.datetime.utcnow()-datetime.timedelta(days=1),
+                                                       datetime.datetime.utcnow()+datetime.timedelta(days=1),
+                                                       "", "CEP4")['task_id']
+        task = self.radb.getTask(task_id)
+        cep_storage_resource = next(r for r in self.radb.getResources(resource_types='storage', include_availability=True) if 'CEP4' in r['name'])
+        self.radb.insertResourceClaim(cep_storage_resource['id'], task_id, task['starttime'], task['endtime'],
+                                      0.75*cep_storage_resource['total_capacity'], "", 0)
+        self.radb.updateTaskAndResourceClaims(task_id, claim_status='claimed', task_status='scheduled')
+        self.assertEqual('scheduled', self.radb.getTask(task_id)['status'])
+
+        # make sure the estimater mock asks for too much storage which wont fit during scheduling
+        def rerpc_mock_get_estimated_resources(specification_tree):
+            otdb_id = specification_tree['otdb_id']
+            estimates = deepcopy(self.rerpc_replymessage[str(otdb_id)])
+            estimates['estimates'][0]['resource_types']['storage'] = 0.75*cep_storage_resource['total_capacity']
+            return estimates
+        self.rerpc_mock.get_estimated_resources.side_effect = rerpc_mock_get_estimated_resources
 
+        # now test the resource_assigner.do_assignment. Should not succeed. Task and claims should go to conflict status.
+        self.specification_tree['status'] = 'prescheduled'
         self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
 
-        self.spec_mock.read_from_radb.assert_called_with(self.spec_mock.radb_id)
-
-    def test_do_assignment_should_set_status_to_conflict_again_when_cant_schedule_and_in_conflict(self):
-        self.spec_mock.status = 'conflict'
-        self.dwell_scheduler_mock().allocate_resources.return_value = False
+        # check if task is in the radb, and if status is in conflict
+        resulting_task = self.radb.getTask(otdb_id=self.specification_tree['otdb_id'])
+        self.assertIsNotNone(resulting_task)
+        self.assertEqual('conflict', resulting_task['status'])
 
-        self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
-
-        self.spec_mock.set_status.assert_called_with('conflict')
+        # check if TaskConflict notification was logged and send
+        content = {'radb_id': resulting_task['id'], 'otdb_id': resulting_task['otdb_id'], 'mom_id': resulting_task['mom_id']}
+        subject = 'TaskConflict'
+        self.assertBusNotificationAndLogging(content, subject)
 
     def test_do_assignment_notifies_bus_when_it_was_unable_to_schedule_Error(self):
-        content = {'radb_id': self.task_id, 'otdb_id': self.task_otdb_id, 'mom_id': self.task_mom_id}
-        subject = 'Task' + 'Error'
-
-        self.spec_mock.status = 'error'
-        self.dwell_scheduler_mock().allocate_resources.return_value = False
+        # make sure the estimater mock asks for more storage than available resulting in TaskError
+        def rerpc_mock_get_estimated_resources(specification_tree):
+            otdb_id = specification_tree['otdb_id']
+            estimates = deepcopy(self.rerpc_replymessage[str(otdb_id)])
+            cep_storage_resource = next(r for r in
+                                        self.radb.getResources(resource_types='storage', include_availability=True)
+                                        if 'CEP4' in r['name'])
+            estimates['estimates'][0]['resource_types']['storage'] = 1+cep_storage_resource['total_capacity']
+            return estimates
+        self.rerpc_mock.get_estimated_resources.side_effect = rerpc_mock_get_estimated_resources
 
+        # now test the resource_assigner.do_assignment. Should not succeed. Task should go to error status.
+        self.specification_tree['status'] = 'prescheduled'
         self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
 
+        # check if task is in the radb, and if status is in error
+        resulting_task = self.radb.getTask(otdb_id=self.specification_tree['otdb_id'])
+        self.assertIsNotNone(resulting_task)
+        self.assertEqual('error', resulting_task['status'])
+
+        # check if TaskError notification was logged and send
+        content = {'radb_id': resulting_task['id'], 'otdb_id': resulting_task['otdb_id'], 'mom_id': resulting_task['mom_id']}
+        subject = 'TaskError'
         self.assertBusNotificationAndLogging(content, subject)
 
     def test_do_assignment_should_set_status_to_error_again_when_cant_schedule_and_not_in_conflict(self):
-        self.spec_mock.status = 'error'
-        self.dwell_scheduler_mock().allocate_resources.return_value = False
+        # make sure the estimater mock asks for more storage than available resulting in TaskError
+        def rerpc_mock_get_estimated_resources(specification_tree):
+            otdb_id = specification_tree['otdb_id']
+            estimates = deepcopy(self.rerpc_replymessage[str(otdb_id)])
+            cep_storage_resource = next(r for r in
+                                        self.radb.getResources(resource_types='storage', include_availability=True)
+                                        if 'CEP4' in r['name'])
+            estimates['estimates'][0]['resource_types']['storage'] = 1+cep_storage_resource['total_capacity']
+            return estimates
+        self.rerpc_mock.get_estimated_resources.side_effect = rerpc_mock_get_estimated_resources
 
-        self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
+        # check if the task assignment results in an error twice (apparently it didn't someday for whatever reason)
+        for i in range(2):
+            self.specification_tree['status'] = 'prescheduled'
+            self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
 
-        self.spec_mock.set_status.assert_called_with('error')
+            # check if task is in the radb, and if status is in error
+            resulting_task = self.radb.getTask(otdb_id=self.specification_tree['otdb_id'])
+            self.assertIsNotNone(resulting_task)
+            self.assertEqual('error', resulting_task['status'])
 
-    def test_do_assignment_logs_task_data_removal_if_task_is_pipeline(self):
-        self.sqrpc_mock.getDiskUsageForOTDBId.return_value = {'found': True, 'disk_usage': 10}
+            # check if TaskError notification was logged and send
+            content = {'radb_id': resulting_task['id'], 'otdb_id': resulting_task['otdb_id'], 'mom_id': resulting_task['mom_id']}
+            subject = 'TaskError'
+            self.assertBusNotificationAndLogging(content, subject)
+
+            # reset the mock calls for next round
+            self.ra_notification_bus_mock().send.reset_mock()
+            self.logger_mock.info.send.reset_mock()
 
+    def test_do_assignment_logs_task_data_removal_if_task_is_pipeline(self):
+        self.specification_tree['status'] = 'prescheduled'
         self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
 
         self.logger_mock.info.assert_any_call("removing data on disk from previous run for otdb_id %s", self.otdb_id)
 
     def test_do_assignment_removes_task_data_if_task_is_pipeline(self):
-        self.sqrpc_mock.getDiskUsageForOTDBId.return_value = {'found': True, 'disk_usage': 10}
-
+        self.specification_tree['status'] = 'prescheduled'
         self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
 
-        self.curpc_mock.removeTaskData.assert_any_call(self.task_otdb_id)
+        self.curpc_mock.removeTaskData.assert_any_call(self.specification_tree['otdb_id'])
 
     def test_do_assignment_logs_when_taks_data_could_not_be_deleted(self):
         message = "file was locked"
-        self.sqrpc_mock.getDiskUsageForOTDBId.return_value = {'found': True, 'disk_usage': 10}
         self.curpc_mock.removeTaskData.return_value = {'deleted': False, 'message': message}
 
+        self.specification_tree['status'] = 'prescheduled'
         self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
 
         self.logger_mock.warning.assert_any_call(
@@ -1936,26 +705,42 @@ class ResourceAssignerTest(unittest.TestCase):
 
     def test_do_assignment_logs_exception_from_curcp_removeTaskData(self):
         exception_str = "Error something went wrong"
-        self.sqrpc_mock.getDiskUsageForOTDBId.return_value = {'found': True, 'disk_usage': 10}
         self.curpc_mock.removeTaskData.side_effect = Exception(exception_str)
 
-        self.resource_assigner.do_assignment(self.specification_tree['otdb_id'],
-                                             self.specification_tree)
+        otdb_id = self.specification_tree['otdb_id']
+        # assure task is not known yet
+        self.assertIsNone(self.radb.getTask(otdb_id=otdb_id))
+
+        self.specification_tree['status'] = 'prescheduled'
+        self.resource_assigner.do_assignment(otdb_id, self.specification_tree)
 
         self.logger_mock.error.assert_any_call("Exception in cleaning up earlier data: %s", exception_str)
 
     def test_do_assignment_notifies_bus_when_task_is_scheduled(self):
-        content = {'radb_id': self.task_id, 'otdb_id': self.task_otdb_id, 'mom_id': self.task_mom_id}
-        subject = 'Task' + 'Scheduled'
-
+        self.specification_tree['status'] = 'prescheduled'
         self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
 
+        # check if task is in the radb, and if status is scheduled
+        resulting_task = self.radb.getTask(otdb_id=self.specification_tree['otdb_id'])
+        self.assertIsNotNone(resulting_task)
+        self.assertEqual('scheduled', resulting_task['status'])
+
+        # check if TaskScheduled notification was logged and send
+        content = {'radb_id': resulting_task['id'], 'otdb_id': resulting_task['otdb_id'], 'mom_id': resulting_task['mom_id']}
+        subject = 'TaskScheduled'
         self.assertBusNotificationAndLogging(content, subject)
 
     def test_do_assignement_set_status_on_spec_when_scheduleable(self):
-        self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
+        otdb_id = self.specification_tree['otdb_id']
+        # assure task is not known yet
+        self.assertIsNone(self.radb.getTask(otdb_id=otdb_id))
+
+        self.specification_tree['status'] = 'prescheduled'
+        self.resource_assigner.do_assignment(otdb_id, self.specification_tree)
 
-        self.spec_mock.set_status.assert_called_with('scheduled')
+        # assure task is known now, and scheduled
+        self.assertIsNotNone(self.radb.getTask(otdb_id=otdb_id))
+        self.assertEqual('scheduled', self.radb.getTask(otdb_id=otdb_id)['status'])
 
     def assertBusNotificationAndLogging(self, content, subject):
         self.assertTrue(self.ra_notification_bus_send_called_with(content, "%s.%s" %(DEFAULT_RA_NOTIFICATION_PREFIX, subject)))
@@ -1967,6 +752,7 @@ class ResourceAssignerTest(unittest.TestCase):
         self.otdbrpc_mock.taskSetSpecification.side_effect = Exception(exception_str)
 
         with self.assertRaisesRegexp(Exception, exception_str):
+            self.mom_bug_specification_tree['status'] = 'prescheduled'
             self.resource_assigner.do_assignment(self.mom_bug_specification_tree['otdb_id'],
                                                  self.mom_bug_specification_tree)
 
@@ -2019,14 +805,72 @@ class ResourceAssignerTest(unittest.TestCase):
         self.obscontrol_mock.abort_observation.assert_called_with(spec.otdb_id)
 
     # SW-800 The schedulers need open and close called (using context manager)
-    def test_do_assignement_uses_context_manager_on_dwell_scheduler(self):
-        self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
+    def test_do_assignement_uses_context_manager_on_schedulers(self):
+        with mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.BasicScheduler') as basic_scheduler_mock:
+            with mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.PriorityScheduler') as prio_scheduler_mock:
+                basic_scheduler_mock().allocate_resources.return_value = (False, None)
+                prio_scheduler_mock().allocate_resources.return_value = (False, None)
+
+                self.specification_tree['status'] = 'prescheduled'
+                self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree)
 
-        self.dwell_scheduler_mock().__enter__.assert_called()
-        self.dwell_scheduler_mock().__exit__.assert_called()
+                basic_scheduler_mock().__enter__.assert_called()
+                basic_scheduler_mock().__exit__.assert_called()
+
+                prio_scheduler_mock().__enter__.assert_called()
+                prio_scheduler_mock().__exit__.assert_called()
 
     #This class is currently missing any tests of interaction between tasks already scheduled and new tasks,
     # e.g. triggered ones. It would require a totally different way to set up the tests to be able to test this.
 
+    def test_do_assignment_does_not_raise_on_inserting_predecessors(self):
+        '''SW-816: When scheduling a successor task, it failed on 'duplicate key value violates unique constraint "task_predecessor_unique"' error in the radb.
+        This test proves correct/expected behaviour.'''
+        predecessor_spec = deepcopy(self.specification_tree)
+        predecessor_otdb_id = predecessor_spec['otdb_id']
+        predecessor_spec['status'] = 'prescheduled'
+        predecessor_spec['predecessors'] = []
+        self.resource_assigner.do_assignment(predecessor_otdb_id, predecessor_spec)
+
+        # check if task is in the radb, and if status is scheduled
+        predecessor_task = self.radb.getTask(otdb_id=predecessor_otdb_id)
+        self.assertIsNotNone(predecessor_task)
+        self.assertEqual('scheduled', predecessor_task['status'])
+
+
+        successor_spec = deepcopy(self.specification_tree)
+        successor_spec['otdb_id'] += 1000
+        successor_spec['mom_id'] += 1000
+        successor_otdb_id = successor_spec['otdb_id']
+        successor_spec['status'] = 'prescheduled'
+        successor_spec['predecessors'] = [predecessor_spec]
+
+        # let the mocked resource estimator return the same estimates for this new otdb_id+1000
+        def rerpc_mock_get_estimated_resources(specification_tree):
+            otdb_id = specification_tree['otdb_id']-1000
+            return self.rerpc_replymessage[str(otdb_id)]
+        self.rerpc_mock.get_estimated_resources.side_effect = rerpc_mock_get_estimated_resources
+
+        # let the momrpc_mock provide the proper linkage between the tasks
+        self.momrpc_mock.getPredecessorIds.return_value = {str(successor_spec['mom_id']): [predecessor_spec['mom_id']]}
+        self.momrpc_mock.getSuccessorIds.return_value = {str(predecessor_spec['mom_id']): [successor_spec['mom_id']]}
+
+        # it should be possible to scheduled the successor twice and link it twice to the predecessor.
+        # the second time, it should just be 'cleaned-up' and rescheduled/relinked.
+        for i in range(2):
+            self.resource_assigner.do_assignment(successor_otdb_id, successor_spec)
+
+            # check if task is in the radb, and if status is scheduled
+            successor_task = self.radb.getTask(otdb_id=successor_otdb_id)
+            self.assertIsNotNone(successor_task)
+            self.assertEqual('scheduled', successor_task['status'])
+            self.assertEqual([predecessor_task['id']], successor_task['predecessor_ids'], )
+
+            # check if predecessor_task is also linked to its successor
+            predecessor_task = self.radb.getTask(otdb_id=predecessor_otdb_id)
+            self.assertEqual([successor_task['id']], predecessor_task['successor_ids'], )
+
+
+
 if __name__ == '__main__':
     unittest.main()
diff --git a/SAS/ResourceAssignment/ResourceAssigner/test/t_schedulers.py b/SAS/ResourceAssignment/ResourceAssigner/test/t_schedulers.py
index c03af9a1bfe7d7be28c31594405840211b4bc76d..85efbd6aa7c5be61a49cfda0c5c576349fcdb40e 100755
--- a/SAS/ResourceAssignment/ResourceAssigner/test/t_schedulers.py
+++ b/SAS/ResourceAssignment/ResourceAssigner/test/t_schedulers.py
@@ -37,41 +37,34 @@ from lofar.sas.resourceassignment.database.radb import FETCH_ONE
 import logging
 logger = logging.getLogger(__name__)
 
-try:
-    from . import radb_common_testing
-except (ImportError, SystemError):
-    import radb_common_testing
+from lofar.sas.resourceassignment.database.testing.radb_common_testing import RADBCommonTestMixin
 
-
-def setUpModule():
-    return radb_common_testing.setUpModule()
-
-
-def tearDownModule():
-    return radb_common_testing.tearDownModule()
-
-
-class SchedulerTest(radb_common_testing.RADBCommonTest):
+class SchedulerTest(RADBCommonTestMixin, unittest.TestCase):
     """ create test radb postgres instance, and use that in a ResourceAvailabilityChecker"""
 
-    def setUp(self):
-        super(SchedulerTest, self).setUp()
-        self.resource_availability_checker = ResourceAvailabilityChecker(self.radb)
-        self._enforce_limited_station_group_list()
+    @classmethod
+    def setUpClass(cls):
+        super().setUpClass()
+        cls._enforce_limited_station_group_list()
 
-    def _enforce_limited_station_group_list(self):
+    @classmethod
+    def _enforce_limited_station_group_list(cls):
         # for test simplicity, create a simple virtual instrument which makes debugging easier.
         # this is safe, because we are working on a test database
 
         LIMITED_STATION_GROUP_LIST = ('CS001', 'CS002', 'RS106', 'RS205')
 
-        unwanted_resource_group_ids = [rg['id'] for rg in self.radb.getResourceGroups()
+        unwanted_resource_group_ids = [rg['id'] for rg in cls.radb.getResourceGroups()
                                        if rg['type'] == 'station' and rg['name'] not in LIMITED_STATION_GROUP_LIST]
 
-        self.radb.executeQuery("DELETE FROM virtual_instrument.resource_group rg WHERE rg.id in (%s)" % (
-                                ', '.join([str(id) for id in unwanted_resource_group_ids])),)
-        self.radb.commit()
+        if unwanted_resource_group_ids:
+            cls.radb.executeQuery("DELETE FROM virtual_instrument.resource_group rg WHERE rg.id in (%s)" % (
+                                    ', '.join([str(id) for id in unwanted_resource_group_ids])),)
+            cls.radb.commit()
 
+    def setUp(self):
+        super().setUp()
+        self.resource_availability_checker = ResourceAvailabilityChecker(self.radb)
 
 class BasicSchedulerTest(SchedulerTest):
     def new_task(self, mom_otdb_id=0, starttime=None, endtime=None):
@@ -89,7 +82,7 @@ class BasicSchedulerTest(SchedulerTest):
         if endtime is None:
             endtime = datetime.datetime(2017, 1, 1, 2, 0, 0)
 
-        return self.radb.insertSpecificationAndTask(mom_id=mom_otdb_id,
+        return self.radb.insertOrUpdateSpecificationAndTask(mom_id=mom_otdb_id,
                                                     otdb_id=mom_otdb_id,
                                                     task_status='approved',
                                                     task_type='observation',
@@ -111,7 +104,7 @@ class BasicSchedulerTest(SchedulerTest):
         return BasicScheduler(task_id,
                               specification_tree if specification_tree else self.get_specification_tree(task_id),
                               resource_estimator if resource_estimator else lambda _:[],
-                              self.resource_availability_checker, self.radb.dbcreds)
+                              self.resource_availability_checker, self.radb)
 
     def get_station_bandwidth_max_capacity(self):
         resource_CS001bw0 = [r for r in self.radb.getResources(resource_types="bandwidth", include_availability=True)
@@ -251,7 +244,7 @@ class StationSchedulerTest(BasicSchedulerTest):
         return StationScheduler(task_id,
                                 specification_tree if specification_tree else self.get_specification_tree(task_id),
                                 resource_estimator if resource_estimator else self.fake_resource_estimator,
-                                self.resource_availability_checker, self.radb.dbcreds)
+                                self.resource_availability_checker, self.radb)
 
     def fake_resource_estimator(self, specification_tree):
         """ Return an estimate for each station, plus a fixed storage claim of half the available storage capacity. """
@@ -496,7 +489,7 @@ class PrioritySchedulerTest(StationSchedulerTest):
         self.mock_datetime()
 
     def new_task_without_momid(self, otdb_id):
-        return self.radb.insertSpecificationAndTask(mom_id=None,
+        return self.radb.insertOrUpdateSpecificationAndTask(mom_id=None,
                                                     otdb_id=otdb_id,
                                                     task_status='approved',
                                                     task_type='observation',
@@ -519,7 +512,7 @@ class PrioritySchedulerTest(StationSchedulerTest):
         return PriorityScheduler(task_id,
                                  specification_tree if specification_tree else self.get_specification_tree(task_id),
                                  resource_estimator if resource_estimator else self.fake_resource_estimator,
-                                 self.resource_availability_checker, self.radb.dbcreds)
+                                 self.resource_availability_checker, self.radb)
 
     def test_unschedule_lower_priority_future_task(self):
         """
@@ -910,7 +903,7 @@ class DwellSchedulerTest(PrioritySchedulerTest):
                               min_starttime,
                               max_starttime,
                               datetime.timedelta(hours=1),            # duration
-                              self.resource_availability_checker, self.radb.dbcreds)
+                              self.resource_availability_checker, self.radb)
 
     def test_no_dwell(self):
         """ Whether a task will not dwell unnecessarily on an empty system. """
diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/CMakeLists.txt b/SAS/ResourceAssignment/ResourceAssignmentDatabase/CMakeLists.txt
index 669f6c8fcc4e18824c5f383c33f834499b9639fc..dc5157d4721e30a115e9c35336aed26a68007822 100644
--- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/CMakeLists.txt
+++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/CMakeLists.txt
@@ -10,9 +10,7 @@ set(_py_files
   config.py
   radb.py
   radbpglistener.py
-  radbbuslistener.py
-  tests/radb_common_testing.py
-)
+  radbbuslistener.py)
 
 python_install(${_py_files} DESTINATION lofar/sas/resourceassignment/database)
 
diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb.py b/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb.py
index 9fa72c99b29f066c3efdbfce4ebf764b9dc43527..43a7c7674b9666c27efe62c574796e425ef6e7ff 100644
--- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb.py
+++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb.py
@@ -35,9 +35,14 @@ class RADBError(Exception):
     pass
 
 class RADatabase(PostgresDatabaseConnection):
-    def __init__(self, dbcreds: dbcredentials.DBCredentials,
+    def __init__(self, dbcreds: dbcredentials.DBCredentials=None,
                  num_connect_retries: int=5,
                  connect_retry_interval: float=1.0):
+
+        if dbcreds is None:
+            dbcreds = dbcredentials.DBCredentials().get("RADB")
+            logger.info("Read default RADB dbcreds from disk: %s" % dbcreds.stringWithHiddenPassword())
+
         super().__init__(dbcreds=dbcreds,
                          auto_commit_selects=False,
                          num_connect_retries=num_connect_retries,
@@ -1315,6 +1320,16 @@ class RADatabase(PostgresDatabaseConnection):
             return self._cursor.rowcount > 0
         return True
 
+    def deleteResourceClaimForTask(self, task_id, commit=True):
+        query = '''LOCK TABLE resource_allocation.resource_claim, resource_allocation.resource_usage, resource_allocation.task IN EXCLUSIVE MODE; '''\
+                '''DELETE FROM resource_allocation.resource_claim
+                   WHERE resource_allocation.resource_claim.task_id = %s;'''
+
+        self.executeQuery(query, (task_id,))
+        if commit:
+            self.commit()
+        return self._cursor.rowcount > 0
+
     def updateResourceClaim(self, resource_claim_id, resource_id=None, task_id=None, starttime=None, endtime=None,
                             status=None, claim_size=None, username=None, used_rcus=None, user_id=None,
                             commit=True):
@@ -1552,41 +1567,31 @@ class RADatabase(PostgresDatabaseConnection):
         else:
             self.executeQuery('SELECT * from resource_allocation.rebuild_resource_usages_from_claims_for_resource_of_status(%s, %s)', (resource_id, claim_status_id), fetch=FETCH_NONE)
 
-    def insertSpecificationAndTask(self, mom_id, otdb_id, task_status, task_type, starttime, endtime, content, cluster, commit=True):
+    def insertOrUpdateSpecificationAndTask(self, mom_id, otdb_id, task_status, task_type, starttime, endtime, content, cluster, commit=True):
         '''
         Insert a new specification and task in one transaction.
-        Removes existing task with same otdb_id if present in the same transaction.
-        Removes existing task with same mom_id if present in the same transaction.
+        Removes resource_claims for existing task with same otdb_id if present in the same transaction.
         '''
         try:
             existing_task = self.getTask(otdb_id=otdb_id)
-            if existing_task:
-                # delete old specification, task, and resource claims using cascaded delete
-                self.deleteSpecification(existing_task['specification_id'], False)
-            else:
-                if mom_id:
-                    existing_task = self.getTask(mom_id=mom_id)
-                    if existing_task:
-                        # delete old specification, task, and resource claims using cascaded delete
-                        self.deleteSpecification(existing_task['specification_id'], False)
-
-            specId = self.insertSpecification(starttime, endtime, content, cluster, False)
-            taskId = self.insertTask(mom_id, otdb_id, task_status, task_type, specId, False)
-
-            if specId >= 0 and taskId >= 0:
-                # restore "old" predecessor/successor relationships if needed
-                if existing_task:
-                    if existing_task['predecessor_ids']:
-                        self.insertTaskPredecessors(taskId, existing_task['predecessor_ids'], False)
-                    if existing_task['successor_ids']:
-                        for suc_id in existing_task['successor_ids']:
-                            self.insertTaskPredecessor(suc_id, taskId, False)
-
-                if commit:
-                    self.commit()
-                return {'inserted': True, 'specification_id': specId, 'task_id': taskId}
+            if existing_task is None and mom_id is not None:
+                existing_task = self.getTask(mom_id=mom_id)
+
+            if existing_task is not None:
+                # delete any existing resource_claims and update the properties of the spec and task
+                logger.info("insertOrUpdateSpecificationAndTask: a task with the same mom/otdb id is already known: existing_task=%s. Clearing current resource claims (if any) and updating specification and task.", existing_task)
+                specId = existing_task['specification_id']
+                taskId = existing_task['id']
+                self.deleteResourceClaimForTask(existing_task['id'], False)
+                self.updateSpecification(specId, starttime=starttime, endtime=endtime, content=content, cluster=cluster, commit=False)
+                self.updateTask(taskId, mom_id=mom_id, otdb_id=otdb_id, task_status=task_status, task_type=task_type, commit=False)
             else:
-                self.rollback()
+                specId = self.insertSpecification(starttime, endtime, content, cluster, False)
+                taskId = self.insertTask(mom_id, otdb_id, task_status, task_type, specId, False)
+
+            if commit:
+                self.commit()
+            return {'inserted': True, 'specification_id': specId, 'task_id': taskId}
         except Exception as e:
             logger.error(e)
             self.rollback()
diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb/sql/add_functions_and_triggers.sql b/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb/sql/add_functions_and_triggers.sql
index e89ea38d82e0b932e5403cf4dc27360b2d828b9c..a24822796b807ce3f5963758bd38dd0034a77566 100644
--- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb/sql/add_functions_and_triggers.sql
+++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb/sql/add_functions_and_triggers.sql
@@ -916,5 +916,31 @@ CREATE TRIGGER T_after_claim_insertupdatedelete
 
 ---------------------------------------------------------------------------------------------------------------------
 
+CREATE OR REPLACE FUNCTION resource_allocation.after_claim_truncate()
+  RETURNS trigger AS
+$BODY$
+DECLARE
+BEGIN
+    TRUNCATE resource_allocation.resource_usage CASCADE;
+    TRUNCATE resource_allocation.resource_usage_delta CASCADE;
+    RETURN NEW;
+END;
+$BODY$
+  LANGUAGE plpgsql VOLATILE
+  COST 100;
+ALTER FUNCTION resource_allocation.after_claim_truncate()
+  OWNER TO resourceassignment;
+COMMENT ON FUNCTION resource_allocation.after_claim_truncate()
+  IS 'tables resource_allocation.resource_usage and resource_allocation.resource_usage do not have references to the resource_allocation.resource_claim table. So there is no automatic cascading truncate. This function and the truncate trigger on resource_allocation.resource_claim makes sure these tables are truncated as well.';
+
+DROP TRIGGER IF EXISTS T_after_claim_truncate ON resource_allocation.resource_claim CASCADE;
+CREATE TRIGGER T_after_claim_truncate
+  AFTER TRUNCATE
+  ON resource_allocation.resource_claim
+  FOR EACH STATEMENT
+  EXECUTE PROCEDURE resource_allocation.after_claim_truncate();
+
+---------------------------------------------------------------------------------------------------------------------
+
 
 COMMIT;
diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/CMakeLists.txt b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/CMakeLists.txt
index cd4293dd4bc5d3cee8357e9cbb747ce9bb24058c..ccac9fb675a0aa71459daa90738baa0ccb2b9256 100644
--- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/CMakeLists.txt
+++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/CMakeLists.txt
@@ -1,12 +1,18 @@
 # $Id: CMakeLists.txt 32679 2015-10-26 09:31:56Z schaap $
 include(LofarCTest)
-include(FindPythonModule)
 
-find_python_module(testing.postgresql)
-find_python_module(dateutil)
+if(BUILD_TESTING)
+    include(PythonInstall)
+    include(FindPythonModule)
 
-lofar_add_test(t_radb_functionality)
-lofar_add_test(t_radb_performance)
+    find_python_module(testing.postgresql)
+    find_python_module(dateutil)
 
-set_tests_properties(t_radb_functionality PROPERTIES TIMEOUT 300)
-set_tests_properties(t_radb_performance PROPERTIES TIMEOUT 300)
+    python_install(radb_common_testing.py DESTINATION lofar/sas/resourceassignment/database/testing)
+
+    lofar_add_test(t_radb_functionality)
+    lofar_add_test(t_radb_performance)
+
+    set_tests_properties(t_radb_functionality PROPERTIES TIMEOUT 300)
+    set_tests_properties(t_radb_performance PROPERTIES TIMEOUT 300)
+endif()
\ No newline at end of file
diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/radb_common_testing.py b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/radb_common_testing.py
index 7487aed9b10d20fb1bcf255bcb960e68e3180e0f..e1d89daaf30d188d0da93daa25011c5c9e87802c 100755
--- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/radb_common_testing.py
+++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/radb_common_testing.py
@@ -29,84 +29,76 @@ logger = logging.getLogger(__name__)
 
 import testing.postgresql
 from lofar.common.dbcredentials import Credentials
+from lofar.common.postgres import PostgresDatabaseConnection, FETCH_ALL
 from lofar.sas.resourceassignment.database.radb import RADatabase
 
-# Create shared test database for better performance
-database_credentials = None
-Postgresql = None
+class RADBCommonTestMixin():
+    '''
+    A common test mixin class from which you can derive to get a freshly setup postgres testing instance with the latest RADB sql setup scripts applied.
+    '''
 
-def setUpModule():
-    global database_credentials, Postgresql
-    database_credentials = Credentials()
-    Postgresql = testing.postgresql.PostgresqlFactory(cache_initialized_db = True)
-
-def tearDownModule():
-    # clear cached database at end of tests
-    logger.info('tearDownModule')
-    Postgresql.clear_cache()
-
-class RADBCommonTest(unittest.TestCase):
-
-    def setUp(self):
-        logger.info('setting up test RA database...')
+    @classmethod
+    def setUpClass(cls):
+        logger.info('setting up test database instance...')
         # connect to shared test db
-        self.postgresql = Postgresql()    # fresh db instead of shared one: self.postgresql = testing.postgresql.Postgresql()
-
-        # set up fixtures
-        # Note: In theory, this can be moved to the PostgresqlFactory call as kwarg 'on_initialized=populatedb'
-        # ...but for some reason that was much slower than keeping it here.
-        self._setup_database()
+        cls.postgresql = testing.postgresql.PostgresqlFactory(cache_initialized_db=True)()
+        cls.dbcreds    = Credentials()
 
         # update credentials (e.g. port changes for each test)
-        database_credentials.host = self.postgresql.dsn()['host']
-        database_credentials.database = self.postgresql.dsn()['database']
-        database_credentials.port = self.postgresql.dsn()['port']
-
-        # connect with useradministration role for tests
-        self.connection = psycopg2.connect(host = database_credentials.host,
-                                           user = database_credentials.user,
-                                           password = database_credentials.password,
-                                           dbname = database_credentials.database,
-                                           port = database_credentials.port)
-
-        # set up radb python module
-        self.radb = RADatabase(database_credentials)
-        self.radb.connect()
-        logger.info('...finished setting up test RA database')
-
-    def tearDown(self):
-        self.radb.disconnect()
-
-        db_log_file_name = os.path.join(self.postgresql.base_dir, '%s.log' % self.postgresql.name)
-        logger.info('Printing test-postgress-database server log: %s', db_log_file_name)
-        with open(db_log_file_name, 'r') as db_log_file:
-            for line in db_log_file.readlines():
-                print("  postgres log: %s" % line.strip(), file=sys.stderr)
-
-        logger.info('removing test RA database...')
-        self.connection.close()
-
-        # self.Postgresql.clear_cache() # for fresh db during setUp, do instead:
-        self.postgresql.stop()
-
-    def _setup_database(self):
+        cls.dbcreds.host = cls.postgresql.dsn()['host']
+        cls.dbcreds.database = cls.postgresql.dsn()['database']
+        cls.dbcreds.port = cls.postgresql.dsn()['port']
 
         # connect to db as root
-        conn = psycopg2.connect(**self.postgresql.dsn())
+        conn = psycopg2.connect(**cls.postgresql.dsn())
         cursor = conn.cursor()
 
         # set credentials to be used during tests
-        database_credentials.user = 'resourceassignment'
-        database_credentials.password = 'blabla'    # cannot be empty...
+        cls.dbcreds.user = 'resourceassignment'
+        cls.dbcreds.password = 'secret'    # cannot be empty...
 
         # create user role
         # Note: NOSUPERUSER currently raises "permission denied for schema virtual_instrument"
         # Maybe we want to sort out user creation and proper permissions in the sql scripts?
-        query = "CREATE USER %s WITH SUPERUSER PASSWORD '%s'" % (
-            database_credentials.user,
-            database_credentials.password)
+        query = "CREATE USER %s WITH SUPERUSER PASSWORD '%s'" % (cls.dbcreds.user, cls.dbcreds.password)
         cursor.execute(query)
 
+        cursor.close()
+        conn.commit()
+        conn.close()
+
+        logger.info('Finished setting up test database instance. It is avaiblable at: %s', cls.dbcreds.stringWithHiddenPassword())
+
+        cls.radb = RADatabase(cls.dbcreds)
+        cls.radb.connect()
+
+        # set up a fresh copy of the RADB sql schema
+        cls._setup_database(cls.radb)
+
+    def setUp(self):
+        # wipe all tables by truncating specification which cascades into the rest.
+        logger.debug("setUp: Wiping radb tables for each unittest.")
+        self.radb.executeQuery("TRUNCATE TABLE resource_allocation.specification CASCADE;")
+        self.radb.commit()
+
+    @classmethod
+    def tearDownClass(cls):
+        cls.radb.disconnect()
+
+        db_log_file_name = os.path.join(cls.postgresql.base_dir, '%s.log' % cls.postgresql.name)
+        logger.info('Printing test-postgress-database server log: %s', db_log_file_name)
+        with open(db_log_file_name, 'r') as db_log_file:
+            for line in db_log_file.readlines():
+                print("  postgres log: %s" % line.strip(), file=sys.stderr)
+
+        logger.info('removing test RA database at %s', cls.dbcreds.stringWithHiddenPassword())
+        cls.postgresql.stop()
+        logger.info('test RA removed')
+
+    @staticmethod
+    def _setup_database(db: PostgresDatabaseConnection):
+        logger.info('applying RADB sql schema to %s', db)
+
         # populate db tables
         # These are applied in given order to set up test db
         # Note: cannot use create_and_populate_database.sql since '\i' is not understood by cursor.execute()
@@ -120,50 +112,34 @@ class RADBCommonTest(unittest.TestCase):
         for sql_path in sql_createdb_paths:
             logger.debug("setting up database. applying sql file: %s", sql_path)
             with open(sql_path) as sql:
-                cursor.execute(sql.read())
-
-        cursor.close()
-        conn.commit()
-        conn.close()
-
-    def _execute_query(self, query, fetch = False):
-        cursor = self.connection.cursor()
-        cursor.execute(query)
-        ret = None
-        if fetch:
-            ret = cursor.fetchall()
-        cursor.close()
-        self.connection.commit()
-        return ret
-
-    # --- tests start here
-
-    # integrity tests of postgres database itself
-    #
-    # Note: These are meant to make sure the setup generally works and all sql scripts were applied.
-    # I don't see much benefit in full coverage here since it should be all be tested through RADataBase functionality.
-    # Of course new tests can be added here where db functionality like triggers should be tested separately from the
-    # Python part of the job.
+                db.executeQuery(sql.read())
+                db.commit()
 
+class RADBCommonTest(RADBCommonTestMixin, unittest.TestCase):
     # database created?
     def test_select_tables_contains_tables_for_each_schema(self):
-        query = "SELECT table_schema,table_name FROM information_schema.tables"
-        fetch = self._execute_query(query, fetch = True)
-        self.assertTrue('resource_allocation' in str(fetch))
-        self.assertTrue('resource_monitoring' in str(fetch))
-        self.assertTrue('virtual_instrument' in str(fetch))
+        with PostgresDatabaseConnection(self.dbcreds) as connection:
+            query = "SELECT table_schema,table_name FROM information_schema.tables"
+            result = connection.executeQuery(query, fetch=FETCH_ALL)
+            self.assertTrue('resource_allocation' in str(result))
+            self.assertTrue('resource_monitoring' in str(result))
+            self.assertTrue('virtual_instrument' in str(result))
 
     # resource allocation_statics there?
     def test_select_task_types_contains_obervation(self):
-        query = "SELECT * FROM resource_allocation.task_type"
-        fetch = self._execute_query(query, fetch = True)
-        self.assertTrue('observation' in str(fetch))
+        with PostgresDatabaseConnection(self.dbcreds) as connection:
+            query = "SELECT * FROM resource_allocation.task_type"
+            result = connection.executeQuery(query, fetch=FETCH_ALL)
+            self.assertTrue('observation' in str(result))
 
     # virtual instrument there?
     def test_select_virtualinstrument_units_contain_rcuboard(self):
-        query = "SELECT * FROM virtual_instrument.unit"
-        fetch = self._execute_query(query, fetch = True)
-        self.assertTrue('rcu_board' in str(fetch))
+        with PostgresDatabaseConnection(self.dbcreds) as connection:
+            query = "SELECT * FROM virtual_instrument.unit"
+            result = connection.executeQuery(query, fetch=FETCH_ALL)
+            self.assertTrue('rcu_board' in str(result))
+
+__all__ = ['RADBCommonTestMixin']
 
 if __name__ == "__main__":
     os.environ['TZ'] = 'UTC'
diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_functionality.py b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_functionality.py
index 417b891b08286104ba5f3f80f4513d1ac23db780..251bf9abe62045286cbfca4f7faed632ddd1a5f6 100755
--- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_functionality.py
+++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_functionality.py
@@ -31,18 +31,13 @@ logger = logging.getLogger(__name__)
 import unittest.mock as mock
 from multiprocessing import Process, Event
 
-import radb_common_testing
+from lofar.sas.resourceassignment.database.testing.radb_common_testing import RADBCommonTestMixin
 
-from lofar.sas.resourceassignment.database.radb import RADatabase, PostgresDBQueryExecutionError, FETCH_ONE, FETCH_ALL
+from lofar.sas.resourceassignment.database.radb import RADatabase
+from lofar.common.postgres import PostgresDatabaseConnection, PostgresDBQueryExecutionError, FETCH_ONE, FETCH_ALL
 from time import sleep
 
-def setUpModule():
-    return radb_common_testing.setUpModule()
-
-def tearDownModule():
-    return radb_common_testing.tearDownModule()
-
-class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
+class ResourceAssignmentDatabaseTest(RADBCommonTestMixin, unittest.TestCase):
 
     class test_task:
         """ A lot of tests involve manipulation of a task (and its corresponding specification) in the RADB. A test task 
@@ -62,8 +57,10 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
                           cluster='CEP4'):
         query = "INSERT INTO resource_allocation.specification (starttime, endtime, content, cluster) " \
                 "VALUES ('%s', '%s', '%s', '%s') RETURNING id" % (starttime, endtime, content, cluster)
-        res = self._execute_query(query, fetch=True)
-        return res[0][0]
+        with PostgresDatabaseConnection(self.dbcreds) as connection:
+            res = connection.executeQuery(query, fetch=FETCH_ALL)
+            connection.commit()
+            return res[0]['id']
     #
     def test_insert_specification_creates_new_entry(self):
         # insert spec
@@ -72,86 +69,53 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         # check it is there
         query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident
-        fetch = self._execute_query(query, fetch=True)
-        self.assertTrue(content in str(fetch))
+        with PostgresDatabaseConnection(self.dbcreds) as connection:
+            res = connection.executeQuery(query, fetch=FETCH_ALL)
+            self.assertTrue(content in str(res))
 
     def test_update_specification_changes_entry(self):
         # insert spec
         ident = self._insert_test_spec()
 
-        # update existing spec content
-        newcontent = 'testcontent_new'
-        query = "UPDATE resource_allocation.specification SET content = '%s'" % newcontent
-        self._execute_query(query)
+        with PostgresDatabaseConnection(self.dbcreds) as connection:
+            # update existing spec content
+            newcontent = 'testcontent_new'
+            query = "UPDATE resource_allocation.specification SET content = '%s'" % newcontent
+            connection.executeQuery(query)
 
-        # check updated content
-        query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident
-        fetch = self._execute_query(query, fetch=True)
-        self.assertTrue(newcontent in str(fetch))
+            # check updated content
+            query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident
+            res = connection.executeQuery(query, fetch=FETCH_ALL)
+            self.assertTrue(newcontent in str(res))
 
     def test_delete_specification(self):
         # insert spec
         content = 'deletecontent'
         ident = self._insert_test_spec(content=content)
 
-        # make sure it's there
-        query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident
-        fetch = self._execute_query(query, fetch=True)
-        self.assertTrue(content in str(fetch))
+        with PostgresDatabaseConnection(self.dbcreds) as connection:
+            # make sure it's there
+            query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident
+            res = connection.executeQuery(query, fetch=FETCH_ALL)
+            self.assertTrue(content in str(res))
 
-        # delete testspec again
-        query = "DELETE FROM resource_allocation.specification WHERE id = %s" % ident
-        self._execute_query(query)
+            # delete testspec again
+            query = "DELETE FROM resource_allocation.specification WHERE id = %s" % ident
+            connection.executeQuery(query)
 
-        # make sure it's gone
-        query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident
-        fetch = self._execute_query(query, fetch=True)
-        self.assertFalse(content in str(fetch))
+            # make sure it's gone
+            query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident
+            res = connection.executeQuery(query, fetch=FETCH_ALL)
+            self.assertFalse(content in str(res))
 
     # triggers in place?
     def test_insert_specification_swaps_startendtimes_if_needed(self):
         #when inserting spec with start>endtime, should raise error
-        with self.assertRaises(psycopg2.InternalError) as context:
+        with self.assertRaises(PostgresDBQueryExecutionError) as context:
             # insert spec
             starttime = '2017-05-10 12:00:00'
             endtime = '2017-05-10 10:00:00'
-            ident = self._insert_test_spec(starttime=starttime, endtime=endtime)
-
-    # notifications in place?
-    def test_insert_task_triggers_notification(self):
-        # insert specification to not raise INtegrityError
-        ident = self._insert_test_spec()
-
-        # listen on notification
-        cursor = self.connection.cursor()
-        cursor.execute("LISTEN %s;", (psycopg2.extensions.AsIs('task_insert'),))
-
-        # todo: fix this and use this instead to listen for notifications.
-        # todo: ...Problem: For some reason callback function is not called.
-        # set up listener in a way we can check it was called
-        # callback = mock.Mock()
-        # callback.listen.return_value = 42
-        # self.listener.subscribe('task_insert', callback.listen)
-
-        # trigger notification
-        query = "INSERT INTO resource_allocation.task (mom_id, otdb_id, status_id, type_id, specification_id)" \
-                "VALUES (%s, %s, %s, %s, %s)" % (1, 1, 200, 0, ident)
-        self._execute_query(query)
-
-        # wait for notification
-        notification = ''
-        self.connection.poll()
-        while self.connection.notifies:
-            try:
-                notification = self.connection.notifies.pop(0)
-                break
-            except Exception:
-                pass
-
-        self.assertTrue('task_insert' in str(notification))
-
-        # todo: fix listener and use instead of polling:
-        # callback.listen.assert_called()
+            self._insert_test_spec(starttime=starttime, endtime=endtime)
 
     #
     # radb functionality tests
@@ -168,7 +132,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         :returns 2-tuple (task_id, spec_id) or None if task wasn't inserted
         """
 
-        task = self.radb.insertSpecificationAndTask(mom_id=mom_id,
+        task = self.radb.insertOrUpdateSpecificationAndTask(mom_id=mom_id,
                                                     otdb_id=otdb_id,
                                                     task_status=self.test_task.task_status,
                                                     task_type=self.test_task.task_type,
@@ -461,7 +425,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         task_id = self.radb.insertTask(mom_id, otdb_id, 'conflict', 'observation', spec_id)
 
-        self.assertEqual(task_id, 1)
+        self.assertIsNotNone(task_id)
 
     def test_insertTask_duplicate_mom_ids_fails(self):
         """ Verify if radb.insertTask() raises exception when called with already occupied mom_id """
@@ -537,6 +501,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         task_id = self.radb.insertTask(sample_task['mom_id'], sample_task['otdb_id'], sample_task['status'],
                                        sample_task['type'], sample_task['specification_id'])
+        sample_task['id'] = task_id
         task = self.radb.getTask(id=task_id)
 
         self.assertEqual(task, sample_task)
@@ -831,6 +796,25 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         self.assertIs(len(ids), 2)
 
+    def test_reinsert_task_with_predecessor(self):
+        """ Verify if radb.insertTaskPredecessor() returns an ID when called with valid and existing task and
+        predecessor IDs. """
+
+        # Shoot 2 unique tasks and corresponding specifications into RADB
+        task_id_a, _ = self._insert_test_task_and_specification(mom_id=1, otdb_id=10)
+        task_id_b, _ = self._insert_test_task_and_specification(mom_id=2, otdb_id=11)
+
+        # link b to predecessor a and check it
+        self.radb.insertTaskPredecessor(task_id_b, task_id_a)
+        self.assertEqual([], self.radb.getTask(task_id_a)['predecessor_ids'])
+        self.assertEqual([task_id_a], self.radb.getTask(task_id_b)['predecessor_ids'])
+
+        # reinsert task b
+        # check if b is still linked to a
+        task_id_b, _ = self._insert_test_task_and_specification(mom_id=2, otdb_id=11)
+        self.assertEqual([], self.radb.getTask(task_id_a)['predecessor_ids'])
+        self.assertEqual([task_id_a], self.radb.getTask(task_id_b)['predecessor_ids'])
+
     def test_getSpecifications_select_all_on_empty_db_succeeds(self):
         """ Verify if radb.getSpecifications() returns an empty list on an empty RADB """
 
@@ -885,7 +869,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         now = datetime.utcnow()
         now -= timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond) # round to full hour
 
-        result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', now, now+timedelta(hours=1),
+        result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', now, now+timedelta(hours=1),
         'foo', 'CEP4')
         self.assertTrue(result['inserted'])
         spec_id1 = result['specification_id']
@@ -977,7 +961,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         logger.info('-- now test with a 2nd task, and test resource availability, conflicts etc. --')
 
         # another task, fully overlapping with task1
-        result = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', now, now+timedelta(hours=1), 'foo', 'CEP4')
+        result = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', now, now+timedelta(hours=1), 'foo', 'CEP4')
         self.assertTrue(result['inserted'])
         spec_id2 = result['specification_id']
         task_id2 = result['task_id']
@@ -1102,7 +1086,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         task2 = self.radb.getTask(task_id2)
 
         # another task, partially overlapping with both task1 & task3
-        result = self.radb.insertSpecificationAndTask(2, 2, 'approved', 'observation',
+        result = self.radb.insertOrUpdateSpecificationAndTask(2, 2, 'approved', 'observation',
                                                       task1['starttime'] + (task1['endtime']-task1['starttime'])/2,
                                                       task2['starttime'] + (task2['endtime']-task2['starttime'])/2,
                                                       'foo', 'CEP4')
@@ -1187,7 +1171,8 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         # suppose the resource_usages table is broken for some reason, fix it....
         # break it first...
-        self._execute_query('TRUNCATE TABLE resource_allocation.resource_usage;')
+        self.radb.executeQuery('TRUNCATE TABLE resource_allocation.resource_usage;')
+        self.radb.commit()
         #check that it's broken
         self.assertNotEqual(40, self.radb.get_max_resource_usage_between(cep4_id, task1['starttime'], task1['starttime'], 'claimed')['usage'])
         #fix it
@@ -1209,7 +1194,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         future = now + timedelta(hours=2)
 
         # insert one task, and reuse that for multiple  claims
-        result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1),
+        result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1),
                                                       'content', 'CEP4')
         self.assertTrue(result['inserted'])
         task_id = result['task_id']
@@ -1406,7 +1391,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         future = now + timedelta(hours=2)
 
         #insert one task, and reuse that for multiple overlapping claims
-        result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', now, now+timedelta(hours=1),
+        result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', now, now+timedelta(hours=1),
                                                       'foo', 'CEP4')
         self.assertTrue(result['inserted'])
         task_id = result['task_id']
@@ -1528,7 +1513,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         future = now + timedelta(hours=2)
 
         # insert one task, and reuse that for multiple overlapping claims
-        result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'first content',
+        result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'first content',
                                                       'CEP4')
         self.assertTrue(result['inserted'])
         task_id = result['task_id']
@@ -1538,11 +1523,11 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         self.assertEqual(task_id, task['id'])
         self.assertEqual('first content', self.radb.getSpecification(task['specification_id'])['content'])
 
-        # prove that we can re-insert the spec/task, and that the new task is indeed inserted and new
-        result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'second content',
+        # prove that we can re-insert the spec/task, and that the new task is indeed updated
+        result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'second content',
                                                       'CEP4')
         self.assertTrue(result['inserted'])
-        self.assertNotEqual(task_id, result['task_id']) # we should have a new id because it was re-inserted
+        self.assertEqual(task_id, result['task_id']) # as per 20190916 inserting a task again should not yield a new task id.
         task_id = result['task_id']
 
         task = self.radb.getTask(task_id)
@@ -1576,10 +1561,10 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         # prove again that we can re-insert the spec/task (future with claims), and that the new task is indeed inserted and new,
         # and that the claim(s) and usage(s) were actually deleted (via cascading deletes)
-        result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'third content',
+        result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'third content',
                                                       'CEP4')
         self.assertTrue(result['inserted'])
-        self.assertNotEqual(task_id, result['task_id']) # we should have a new id because it was re-inserted
+        self.assertEqual(task_id, result['task_id']) # as per 20190916 inserting a task again should not yield a new task id.
         task_id = result['task_id']
 
         task = self.radb.getTask(task_id)
@@ -1629,10 +1614,10 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         # and prove again that we can re-insert the spec/task (future with claims and a corrupted usage table), and that the new task is indeed inserted and new,
         # and that the claim(s) and usage(s) were actually deleted (via cascading deletes)
         # 2017-08-29: YEAH! the insert fails just like on production. Now we can start making a fix!
-        result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'fourth content',
+        result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'fourth content',
                                                       'CEP4')
         self.assertTrue(result['inserted'])
-        self.assertNotEqual(task_id, result['task_id']) # we should have a new id because it was re-inserted
+        self.assertEqual(task_id, result['task_id']) # as per 20190916 inserting a task again should not yield a new task id.
         task_id = result['task_id']
 
         task = self.radb.getTask(task_id)
@@ -1670,7 +1655,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         start = now - timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond) # round to current full hour
 
         #insert a task
-        result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start, start+timedelta(hours=2), 'foo', 'CEP4')
+        result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start, start+timedelta(hours=2), 'foo', 'CEP4')
         self.assertTrue(result['inserted'])
         task_id = result['task_id']
 
@@ -1795,7 +1780,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         for spec in self.radb.getSpecifications():
             self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims
 
-        result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start,
+        result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start,
                                                       start + timedelta(hours=2), 'foo', 'CEP4')
         task1_id = result1['task_id']
 
@@ -1811,7 +1796,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         # claim same
 
-        result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=5),
+        result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=5),
                                                       start + timedelta(hours=2, minutes=5), 'foo', 'CEP4')
         task2_id = result2['task_id']
 
@@ -1837,7 +1822,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         for spec in self.radb.getSpecifications():
             self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims
 
-        result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start,
+        result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start,
                                                       start + timedelta(hours=2), 'foo', 'CEP4')
         task1_id = result1['task_id']
 
@@ -1853,7 +1838,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         # claim same
 
-        result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=5),
+        result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=5),
                                                       start + timedelta(hours=1, minutes=50), 'foo', 'CEP4')
         task2_id = result2['task_id']
 
@@ -1879,7 +1864,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         for spec in self.radb.getSpecifications():
             self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims
 
-        result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start,
+        result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start,
                                                       start + timedelta(hours=2), 'foo', 'CEP4')
         task1_id = result1['task_id']
 
@@ -1895,7 +1880,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         # claim same
 
-        result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=-5),
+        result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=-5),
                                                       start + timedelta(hours=1, minutes=55), 'foo', 'CEP4')
         task2_id = result2['task_id']
 
@@ -1925,7 +1910,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         for spec in self.radb.getSpecifications():
             self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims
 
-        result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start,
+        result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start,
                                                       start + timedelta(hours=2), 'foo', 'CEP4')
         task1_id = result1['task_id']
 
@@ -1941,7 +1926,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         # claim same
 
-        result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=-5),
+        result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=-5),
                                                       start + timedelta(hours=2, minutes=5), 'foo', 'CEP4')
         task2_id = result2['task_id']
 
@@ -1967,7 +1952,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         for spec in self.radb.getSpecifications():
             self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims
 
-        result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start,
+        result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start,
                                                       start + timedelta(hours=2), 'foo', 'CEP4')
         task1_id = result1['task_id']
 
@@ -1983,7 +1968,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         # claim same
 
-        result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start,
+        result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start,
                                                       start + timedelta(hours=2), 'foo', 'CEP4')
         task2_id = result2['task_id']
 
@@ -2009,7 +1994,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         for spec in self.radb.getSpecifications():
             self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims
 
-        result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start,
+        result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start,
                                                       start + timedelta(hours=2), 'foo', 'CEP4')
         task1_id = result1['task_id']
 
@@ -2025,7 +2010,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         # claim same
 
-        result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(hours=3),
+        result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(hours=3),
                                                       start + timedelta(hours=5), 'foo', 'CEP4')
         task2_id = result2['task_id']
 
@@ -2049,7 +2034,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         for spec in self.radb.getSpecifications():
             self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims
 
-        result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start + timedelta(hours=3),
+        result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start + timedelta(hours=3),
                                                       start + timedelta(hours=5), 'foo', 'CEP4')
         task1_id = result1['task_id']
 
@@ -2065,7 +2050,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         # claim same
 
-        result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start,
+        result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start,
                                                       start + timedelta(hours=2), 'foo', 'CEP4')
         task2_id = result2['task_id']
 
@@ -2104,7 +2089,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         resource_max_cap = self.radb.get_resource_claimable_capacity(RESOURCE_ID, base_time, base_time)
 
         # insert the 'low prio' spec, task...
-        spec_task_low = self.radb.insertSpecificationAndTask(1, 1, 'prescheduled', 'observation',
+        spec_task_low = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'prescheduled', 'observation',
                                                              base_time + timedelta(minutes=5),
                                                              base_time + timedelta(minutes=10), 'foo', 'CEP4')
         task_low_id = spec_task_low['task_id']
@@ -2145,7 +2130,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         # overlapping with the beginning of task_low
         # so, the dwellscheduler finds task_low in task_high's higway
         # so, task_low is aborted by the dwellscheduler (later in the code).
-        spec_task_high1 = self.radb.insertSpecificationAndTask(2, 2, 'approved', 'observation',
+        spec_task_high1 = self.radb.insertOrUpdateSpecificationAndTask(2, 2, 'approved', 'observation',
                                                               base_time,
                                                               base_time + timedelta(minutes=7), 'foo', 'CEP4')
         task_high1_id = spec_task_high1['task_id']
@@ -2214,7 +2199,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         now = datetime.utcnow()
         now -= timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond) # round to full hour
 
-        result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation',
+        result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation',
                                                       now, now+timedelta(hours=1), 'foo', 'CEP4')
         self.assertTrue(result['inserted'])
         self.assertIsNotNone(result['task_id'])
@@ -2257,7 +2242,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         now = datetime.utcnow()
         now -= timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond) # round to full hour
 
-        result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation',
+        result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation',
                                                       now, now+timedelta(hours=1), 'foo', 'CEP4')
         self.assertTrue(result['inserted'])
         self.assertIsNotNone(result['task_id'])
@@ -2332,7 +2317,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         now = now - timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond)
         now = now + timedelta(hours=1)
 
-        spec_task = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation',
+        spec_task = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation',
                                                          now, now + timedelta(minutes=10),
                                                          'foo', 'CEP4')
 
@@ -2405,7 +2390,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         now = datetime.utcnow()
 
-        spec_task = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation',
+        spec_task = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation',
                                                          now, now,  # tasks can have zero duration
                                                          'foo', 'CEP4')
 
@@ -2454,7 +2439,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         resource_max_cap = self.radb.get_resource_claimable_capacity(RESOURCE_ID, base_time, base_time)
 
         # insert a first task and full claim on a resource...
-        spec_task1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation',
+        spec_task1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation',
                                                           base_time + timedelta(minutes=+0),
                                                           base_time + timedelta(minutes=+10), 'foo', 'CEP4')
         self.assertTrue(spec_task1['inserted'])
@@ -2471,7 +2456,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
 
         # insert second (partially overlapping) task and claim on same resource, which we expect to get a conflict status
         # because the first claim already claims the resource fully.
-        spec_task2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation',
+        spec_task2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation',
                                                           base_time + timedelta(minutes=+5),
                                                           base_time + timedelta(minutes=+15), 'foo', 'CEP4')
         self.assertTrue(spec_task2['inserted'])
@@ -2526,7 +2511,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
                                           microseconds=base_time.microsecond)
 
         # insert a first task and full claim on a resource...
-        spec_task = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation',
+        spec_task = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation',
                                                           base_time + timedelta(minutes=-20),
                                                           base_time + timedelta(minutes=-10), 'foo', 'CEP4')
         self.assertTrue(spec_task['inserted'])
@@ -2583,7 +2568,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         RESOURCE_ID = 0
         resource_max_cap = int(self.radb.get_resource_claimable_capacity(RESOURCE_ID, now, now))
 
-        task1_id = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation',
+        task1_id = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation',
                                                         now+timedelta(hours=1),
                                                         now + timedelta(hours=2),
                                                         'content', 'CEP4')['task_id']
@@ -2629,7 +2614,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
                          self.radb.getResourceUsages(task1['starttime'], task1['endtime'], RESOURCE_ID)[RESOURCE_ID]['claimed'])
 
         # insert second task after the first one (not overlapping)
-        task2_id = self.radb.insertSpecificationAndTask(2, 2, 'approved', 'observation',
+        task2_id = self.radb.insertOrUpdateSpecificationAndTask(2, 2, 'approved', 'observation',
                                                         now + timedelta(hours=3),
                                                         now + timedelta(hours=4),
                                                         'content', 'CEP4')['task_id']
@@ -2680,9 +2665,9 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         now = datetime.utcnow()
         now -= timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond)  # round to full hour
         for i in [1,2]:
-            task_id = self.radb.insertSpecificationAndTask(i, i, 'approved', 'observation',
-                                                           now+timedelta(hours=1), now + timedelta(hours=2),
-                                                           'content', 'CEP4')['task_id']
+            task_id = self.radb.insertOrUpdateSpecificationAndTask(i, i, 'approved', 'observation',
+                                                                   now+timedelta(hours=1), now + timedelta(hours=2),
+                                                                   'content', 'CEP4')['task_id']
             task = self.radb.getTask(task_id)
 
             claim = {'resource_id': 0,
@@ -2737,10 +2722,6 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
                         logger.error("exiting updateTask2StartStopTimesLoop because of concurrency issue in updateTask1StatusLoop")
                         return
 
-        # updateTask1StatusLoop()
-        # updateTask2StartStopTimesLoop()
-        # return
-
         # use multiprocessing.Process instead of threading.Thread for real concurrency
         p1 = Process(target=updateTask1StatusLoop, daemon=True)
         p2 = Process(target=updateTask2StartStopTimesLoop, daemon=True)
@@ -2766,7 +2747,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
         # we need a task....
         now = datetime.utcnow()
         now -= timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond)  # round to full hour
-        task_id = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'reservation',
+        task_id = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'reservation',
                                                now + timedelta(hours=1), now + timedelta(hours=2),
                                                'content', 'CEP4')['task_id']
 
diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_performance.py b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_performance.py
index 13ade9788c9c00d4558f80c6733fe60c4176e165..91f1a8cec6576e99706168e308a37a8340fef738 100755
--- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_performance.py
+++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_performance.py
@@ -30,16 +30,10 @@ import logging
 logger = logging.getLogger(__name__)
 
 
-import radb_common_testing
+from lofar.sas.resourceassignment.database.testing.radb_common_testing import RADBCommonTestMixin
 from lofar.sas.resourceassignment.database.radb import RADatabase, FETCH_ONE
 
-def setUpModule():
-    return radb_common_testing.setUpModule()
-
-def tearDownModule():
-    return radb_common_testing.tearDownModule()
-
-class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
+class ResourceAssignmentDatabaseTest(RADBCommonTestMixin, unittest.TestCase):
     def test_resource_usages_performance(self):
         ELAPSED_TRESHOLD = 2.0 #max allowed insert/update/delete time in seconds
 
@@ -77,7 +71,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest):
                         logger.info('starting task and claim scheduling: counter=%s num_claims_per_resource=%s num_claims_to_insert=%s oversubscription_factor=%s',
                                     counter, num_claims_per_resource, num_claims_to_insert, oversubscription_factor)
 
-                        result = self.radb.insertSpecificationAndTask(counter, counter, 'approved', 'observation',
+                        result = self.radb.insertOrUpdateSpecificationAndTask(counter, counter, 'approved', 'observation',
                                                                       now+timedelta(hours=3*counter),
                                                                       now + timedelta(hours=3*counter + 1),
                                                                       'content', 'CEP4')
diff --git a/SAS/ResourceAssignment/ResourceAssignmentService/rpc.py b/SAS/ResourceAssignment/ResourceAssignmentService/rpc.py
index 6307d9cb4f2eb8b2244446a31e56b68f81363bf7..2e2f67b91c069b9d22fd81f3212106e539b6ba09 100644
--- a/SAS/ResourceAssignment/ResourceAssignmentService/rpc.py
+++ b/SAS/ResourceAssignment/ResourceAssignmentService/rpc.py
@@ -240,8 +240,8 @@ class RADBRPC(RPCClientContextManagerMixin):
     def getSpecification(self, id):
         return self._rpc_client.execute('GetSpecification', id=id)
 
-    def insertSpecificationAndTask(self, mom_id, otdb_id, task_status, task_type, starttime, endtime, content, cluster):
-        return self._rpc_client.execute('InsertSpecificationAndTask',
+    def insertOrUpdateSpecificationAndTask(self, mom_id, otdb_id, task_status, task_type, starttime, endtime, content, cluster):
+        return self._rpc_client.execute('insertOrUpdateSpecificationAndTask',
                         mom_id=mom_id,
                         otdb_id=otdb_id,
                         task_status=task_status,
diff --git a/SAS/ResourceAssignment/ResourceAssignmentService/service.py b/SAS/ResourceAssignment/ResourceAssignmentService/service.py
index 888753b4fca36fb8c4d0003b68c330b31fe5b6cc..ddfd4fad090f93ddc6a54e1c870c1a2fb7b4da3c 100644
--- a/SAS/ResourceAssignment/ResourceAssignmentService/service.py
+++ b/SAS/ResourceAssignment/ResourceAssignmentService/service.py
@@ -60,7 +60,7 @@ class RADBServiceMessageHandler(ServiceMessageHandler):
         self.register_service_method('GetTaskTypes', self._getTaskTypes)
         self.register_service_method('GetSpecifications', self._getSpecifications)
         self.register_service_method('GetSpecification', self._getSpecification)
-        self.register_service_method('InsertSpecificationAndTask', self._insertSpecificationAndTask)
+        self.register_service_method('insertOrUpdateSpecificationAndTask', self._insertOrUpdateSpecificationAndTask)
         self.register_service_method('InsertSpecification', self._insertSpecification)
         self.register_service_method('DeleteSpecification', self._deleteSpecification)
         self.register_service_method('UpdateSpecification', self._updateSpecification)
@@ -317,9 +317,9 @@ class RADBServiceMessageHandler(ServiceMessageHandler):
         specification = self.radb.getSpecification(kwargs['id'])
         return specification
 
-    def _insertSpecificationAndTask(self, **kwargs):
-        logger.info('InsertSpecificationAndTask: %s' % dict({k:v for k,v in list(kwargs.items()) if v != None and k != 'content'}))
-        return self.radb.insertSpecificationAndTask(kwargs['mom_id'],
+    def _insertOrUpdateSpecificationAndTask(self, **kwargs):
+        logger.info('insertOrUpdateSpecificationAndTask: %s' % dict({k:v for k,v in list(kwargs.items()) if v != None and k != 'content'}))
+        return self.radb.insertOrUpdateSpecificationAndTask(kwargs['mom_id'],
                                                     kwargs['otdb_id'],
                                                     kwargs['task_status'],
                                                     kwargs['task_type'],
diff --git a/SAS/ResourceAssignment/TaskPrescheduler/taskprescheduler.py b/SAS/ResourceAssignment/TaskPrescheduler/taskprescheduler.py
index a5e6341fccba4509384dc7323f8e09b0c99cee28..760096dd32daec4b34648fd2aa4bd8825686b38e 100644
--- a/SAS/ResourceAssignment/TaskPrescheduler/taskprescheduler.py
+++ b/SAS/ResourceAssignment/TaskPrescheduler/taskprescheduler.py
@@ -32,7 +32,7 @@ from lofar.sas.otdb.OTDBBusListener import OTDBBusListener, OTDBEventMessageHand
 from lofar.sas.otdb.otdbrpc import OTDBRPC
 from lofar.mom.momqueryservice.momqueryrpc import MoMQueryRPC
 from lofar.sas.resourceassignment.taskprescheduler.cobaltblocksize import CorrelatorSettings, StokesSettings, BlockConstraints, BlockSize
-from lofar.sas.resourceassignment.resourceassignmentservice.rpc import RADBRPC
+from lofar.sas.resourceassignment.database.radb import RADatabase
 from lofar.sas.resourceassignment.common.specification import Specification
 from lofar.sas.resourceassignment.common.specification import OUTPUT_PREFIX
 
@@ -88,18 +88,18 @@ class TaskPrescheduler(OTDBEventMessageHandler):
         super().__init__()
         self.otdbrpc  = OTDBRPC.create(exchange=exchange, broker=broker)
         self.momquery = MoMQueryRPC.create(exchange=exchange, broker=broker)
-        self.radbrpc = RADBRPC.create(exchange=exchange, broker=broker)
+        self.radb = RADatabase()
 
     def start_handling(self):
         self.otdbrpc.open()
         self.momquery.open()
-        self.radbrpc.open()
+        self.radb.connect()
         super().start_handling()
 
     def stop_handling(self):
         self.otdbrpc.close()
         self.momquery.close()
-        self.radbrpc.close()
+        self.radb.disconnect()
         super().stop_handling()
 
     def onObservationApproved(self, treeId, modificationTime):
@@ -125,7 +125,7 @@ class TaskPrescheduler(OTDBEventMessageHandler):
         # We get the parset for all tasks we receive instead of just for the ones with
         # a trigger.
         status = "approved"
-        spec = Specification(self.otdbrpc, self.momquery, self.radbrpc)
+        spec = Specification(self.otdbrpc, self.momquery, self.radb)
         spec.set_status(status)
         spec.read_from_OTDB_with_predecessors(treeId, "otdb", {}) #Now checks predecessors, which theoretically could cause race contitions
         spec.read_from_mom()
@@ -165,7 +165,7 @@ class TaskPrescheduler(OTDBEventMessageHandler):
         except Exception as e:
             logger.exception(e)
             logger.error("Problem setting specification or status in OTDB for otdb_id=%s", otdb_id)
-            self.radbrpc.updateTaskStatusForOtdbId(otdb_id, 'error') # We don't catch an exception if this fails.
+            self.radb.updateTaskStatusForOtdbId(otdb_id, 'error') # We don't catch an exception if this fails.
 
 def main():
     from optparse import OptionParser