Loading src/compute/tests/test_slice.py 0 → 100644 +125 −0 Original line number Diff line number Diff line # Copyright 2022-2023 ETSI TeraFlowSDN - TFS OSG (https://tfs.etsi.org/) # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json, random, uuid from typing import Dict, Tuple from compute.service.rest_server.nbi_plugins.ietf_network_slice.bindings.network_slice_services import ( NetworkSliceServices ) # R1 emulated devices # Port 13-0 is Optical # Port 13-1 is Copper R1_UUID = "ed2388eb-5fb9-5888-a4f4-160267d3e19b" R1_PORT_13_0_UUID_OPTICAL = "20440915-1a6c-5e7b-a80f-b0e0e51f066d" R1_PORT_13_1_UUID_COPPER = "ff900d5d-2ac0-576c-9628-a2d016681f9d" # R2 emulated devices # Port 13-0 is Optical # Port 13-1 is Copper R2_UUID = "49ce0312-1274-523b-97b8-24d0eca2d72d" R2_PORT_13_0_UUID_OPTICAL = "214618cb-b63b-5e66-84c2-45c1c016e5f0" R2_PORT_13_1_UUID_COPPER = "4e0f7fb4-5d22-56ad-a00e-20bffb4860f9" # R3 emulated devices # Port 13-0 is Optical # Port 13-1 is Copper R3_UUID = "3bc8e994-a3b9-5f60-9c77-6608b1d08313" R3_PORT_13_0_UUID_OPTICAL = "da5196f5-d651-5def-ada6-50ed6430279d" R3_PORT_13_1_UUID_COPPER = "43d221fa-5701-5740-a129-502131f5bda2" # R4 emulated devices # Port 13-0 is Optical # Port 13-1 is Copper R4_UUID = "b43e6361-2573-509d-9a88-1793e751b10d" R4_PORT_13_0_UUID_OPTICAL = "241b74a7-8677-595c-ad65-cc9093c1e341" R4_PORT_13_1_UUID_COPPER = "c57abf46-caaf-5954-90cc-1fec0a69330e" node_dict = {R1_PORT_13_1_UUID_COPPER: R1_UUID, R2_PORT_13_1_UUID_COPPER: R2_UUID, R3_PORT_13_1_UUID_COPPER: R3_UUID, R4_PORT_13_1_UUID_COPPER: R4_UUID} list_endpoints = [R1_PORT_13_1_UUID_COPPER, R2_PORT_13_1_UUID_COPPER, R3_PORT_13_1_UUID_COPPER, R4_PORT_13_1_UUID_COPPER] list_availability= [99, 99.9, 99.99, 99.999, 99.9999] list_bw = [10, 40, 50, 100, 150, 200, 400] list_owner = ["Telefonica", "CTTC", "Telenor", "ADVA", "Ubitech", "ATOS"] URL_POST = "/restconf/data/ietf-network-slice-service:ietf-nss/network-slice-services" URL_DELETE = "/restconf/data/ietf-network-slice-service:ietf-nss/network-slice-services/slice-service=" def generate_request(seed: str) -> Tuple[Dict, str]: ns = NetworkSliceServices() # Slice 1 suuid = str(uuid.uuid5(uuid.NAMESPACE_DNS, str(seed))) slice1 = ns.slice_service[suuid] slice1.service_description = "Test slice for OFC 2023 demo" slice1.status().admin_status().status = "Planned" # TODO not yet mapped # SDPS: R1 optical to R3 optical sdps1 = slice1.sdps().sdp while True: ep1_uuid = random.choice(list_endpoints) ep2_uuid = random.choice(list_endpoints) if ep1_uuid != ep2_uuid: break sdps1[ep1_uuid].node_id = node_dict.get(ep1_uuid) sdps1[ep2_uuid].node_id = node_dict.get(ep2_uuid) # Connectivity group: Connection construct and 2 sla constrains: # - Bandwidth # - Availability cg_uuid = str(uuid.uuid4()) cg = slice1.connection_groups().connection_group cg1 = cg[cg_uuid] cc1 = cg1.connectivity_construct[0] cc1.cc_id = 5 p2p = cc1.connectivity_construct_type.p2p() p2p.p2p_sender_sdp = ep1_uuid p2p.p2p_receiver_sdp = ep2_uuid slo_custom = cc1.slo_sle_policy.custom() metric_bounds = slo_custom.service_slo_sle_policy().metric_bounds().metric_bound # SLO Bandwidth slo_bandwidth = metric_bounds["service-slo-two-way-bandwidth"] slo_bandwidth.value_description = "Guaranteed bandwidth" slo_bandwidth.bound = int(random.choice(list_bw)) slo_bandwidth.metric_unit = "Gbps" # SLO Availability slo_availability = metric_bounds["service-slo-availability"] slo_availability.value_description = "Guaranteed availability" slo_availability.metric_unit = "percentage" slo_availability.bound = random.choice(list_availability) json_request = {"data": ns.to_json()} #Last, add name and owner manually list_name_owner = [{"tag-type": "owner", "value": random.choice(list_owner)}] json_request["data"]["ietf-network-slice-service:network-slice-services"]["slice-service"][0]["service-tags"] = list_name_owner return (json_request, suuid) if __name__ == "__main__": request = generate_request(123) print(json.dumps(request[0], sort_keys=True, indent=4)) src/load_generator/load_gen/RequestGenerator.py +2 −10 Original line number Diff line number Diff line Loading @@ -39,14 +39,6 @@ ROUTER_ID = { 'R149': '5.5.5.5', 'R155': '5.5.5.1', 'R199': '5.5.5.6', } VIRTUAL_CIRCUIT = { 'R149': '5.5.5.5', 'R155': '5.5.5.1', 'R199': '5.5.5.6', } class RequestGenerator: Loading Loading @@ -269,8 +261,8 @@ class RequestGenerator: src_device_name = self._device_data[src_device_uuid]['name'] src_endpoint_name = self._device_endpoint_data[src_device_uuid][src_endpoint_uuid]['name'] src_router_id = ROUTER_ID.get(src_device_name) src_router_num = int(re.findall(r'^\D*(\d+)', src_device_name)[0]) src_router_id = ROUTER_ID.get(src_device_name) if src_router_id is None: src_router_id = '10.0.0.{:d}'.format(src_router_num) dst_device_name = self._device_data[dst_device_uuid]['name'] Loading Loading @@ -322,8 +314,8 @@ class RequestGenerator: src_device_name = self._device_data[src_device_uuid]['name'] src_endpoint_name = self._device_endpoint_data[src_device_uuid][src_endpoint_uuid]['name'] src_router_id = ROUTER_ID.get(src_device_name) src_router_num = int(re.findall(r'^\D*(\d+)', src_device_name)[0]) src_router_id = ROUTER_ID.get(src_device_name) if src_router_id is None: src_router_id = '10.0.0.{:d}'.format(src_router_num) src_address_ip = '10.{:d}.{:d}.{:d}'.format(x, y, src_router_num) Loading src/pathcomp/frontend/service/algorithms/tools/ComputeSubServices.py +13 −1 Original line number Diff line number Diff line Loading @@ -67,6 +67,7 @@ def convert_explicit_path_hops_to_connections( prv_res_class : Tuple[Optional[int], Optional[DeviceTypeEnum], Optional[str]] = None, None, None for path_hop in path_hops: LOGGER.debug('path_hop={:s}'.format(str(path_hop))) device_uuid = path_hop['device'] if prv_device_uuid == device_uuid: continue device_tuple = device_dict.get(device_uuid) Loading @@ -74,24 +75,33 @@ def convert_explicit_path_hops_to_connections( _,grpc_device = device_tuple res_class = get_resource_classification(grpc_device, device_dict) if res_class[1] in IGNORED_DEVICE_TYPES: continue LOGGER.debug(' prv_res_class={:s}'.format(str(prv_res_class))) LOGGER.debug(' res_class={:s}'.format(str(res_class))) if res_class[1] in IGNORED_DEVICE_TYPES: LOGGER.debug(' ignored') continue if prv_res_class[0] is None: # path ingress LOGGER.debug(' path ingress') connection_stack.put((main_service_uuid, main_service_type, [path_hop], [])) elif prv_res_class[0] > res_class[0]: # create underlying connection LOGGER.debug(' create underlying connection') connection_uuid = str(uuid.uuid4()) prv_service_type = connection_stack.queue[-1][1] service_type = get_service_type(res_class[1], prv_service_type) connection_stack.put((connection_uuid, service_type, [path_hop], [])) elif prv_res_class[0] == res_class[0]: # same resource group kind LOGGER.debug(' same resource group kind') if prv_res_class[1] == res_class[1] and prv_res_class[2] == res_class[2]: # same device type and device controller: connection continues LOGGER.debug(' connection continues') connection_stack.queue[-1][2].append(path_hop) else: # different device type or device controller: chain connections LOGGER.debug(' chain connections') connection = connection_stack.get() connections.append(connection) connection_stack.queue[-1][3].append(connection[0]) Loading @@ -102,6 +112,7 @@ def convert_explicit_path_hops_to_connections( connection_stack.put((connection_uuid, service_type, [path_hop], [])) elif prv_res_class[0] < res_class[0]: # underlying connection ended LOGGER.debug(' underlying connection ended') connection = connection_stack.get() connections.append(connection) connection_stack.queue[-1][3].append(connection[0]) Loading @@ -113,6 +124,7 @@ def convert_explicit_path_hops_to_connections( prv_res_class = res_class # path egress LOGGER.debug(' path egress') connections.append(connection_stack.get()) LOGGER.debug('connections={:s}'.format(str(connections))) assert connection_stack.empty() Loading src/pathcomp/frontend/service/algorithms/tools/ResourceGroups.py +3 −3 Original line number Diff line number Diff line Loading @@ -33,12 +33,12 @@ DEVICE_TYPE_TO_DEEPNESS = { DeviceTypeEnum.EMULATED_P4_SWITCH.value : 60, DeviceTypeEnum.P4_SWITCH.value : 60, DeviceTypeEnum.EMULATED_MICROWAVE_RADIO_SYSTEM.value : 40, DeviceTypeEnum.MICROWAVE_RADIO_SYSTEM.value : 40, DeviceTypeEnum.EMULATED_XR_CONSTELLATION.value : 40, DeviceTypeEnum.XR_CONSTELLATION.value : 40, DeviceTypeEnum.EMULATED_MICROWAVE_RADIO_SYSTEM.value : 30, DeviceTypeEnum.MICROWAVE_RADIO_SYSTEM.value : 30, DeviceTypeEnum.EMULATED_OPEN_LINE_SYSTEM.value : 30, DeviceTypeEnum.OPEN_LINE_SYSTEM.value : 30, Loading src/service/service/task_scheduler/TaskExecutor.py +25 −4 Original line number Diff line number Diff line Loading @@ -12,23 +12,28 @@ # See the License for the specific language governing permissions and # limitations under the License. import json import json, logging from enum import Enum from typing import TYPE_CHECKING, Any, Dict, Optional, Union from common.method_wrappers.ServiceExceptions import NotFoundException from common.proto.context_pb2 import Connection, ConnectionId, Device, DeviceId, Service, ServiceId from common.proto.context_pb2 import Connection, ConnectionId, Device, DeviceDriverEnum, DeviceId, Service, ServiceId from common.tools.context_queries.Connection import get_connection_by_id from common.tools.context_queries.Device import get_device from common.tools.context_queries.Service import get_service_by_id from common.tools.grpc.Tools import grpc_message_list_to_json_string from common.tools.object_factory.Device import json_device_id from context.client.ContextClient import ContextClient from device.client.DeviceClient import DeviceClient from service.service.service_handler_api.Exceptions import ( UnsatisfiedFilterException, UnsupportedFilterFieldException, UnsupportedFilterFieldValueException) from service.service.service_handler_api.ServiceHandlerFactory import ServiceHandlerFactory, get_service_handler_class from service.service.tools.ObjectKeys import get_connection_key, get_device_key, get_service_key if TYPE_CHECKING: from service.service.service_handler_api._ServiceHandler import _ServiceHandler LOGGER = logging.getLogger(__name__) CacheableObject = Union[Connection, Device, Service] class CacheableObjectType(Enum): Loading Loading @@ -169,5 +174,21 @@ class TaskExecutor: self, connection : Connection, service : Service, **service_handler_settings ) -> '_ServiceHandler': connection_devices = self.get_devices_from_connection(connection, exclude_managed_by_controller=True) service_handler_class = get_service_handler_class(self._service_handler_factory, service, connection_devices) try: service_handler_class = get_service_handler_class( self._service_handler_factory, service, connection_devices) return service_handler_class(service, self, **service_handler_settings) except (UnsatisfiedFilterException, UnsupportedFilterFieldException, UnsupportedFilterFieldValueException): dict_connection_devices = { cd_data.name : (cd_uuid, cd_data.name, { (device_driver, DeviceDriverEnum.Name(device_driver)) for device_driver in cd_data.device_drivers }) for cd_uuid,cd_data in connection_devices.items() } LOGGER.exception( 'Unable to select service handler. service={:s} connection={:s} connection_devices={:s}'.format( grpc_message_list_to_json_string(service), grpc_message_list_to_json_string(connection), str(dict_connection_devices) ) ) Loading
src/compute/tests/test_slice.py 0 → 100644 +125 −0 Original line number Diff line number Diff line # Copyright 2022-2023 ETSI TeraFlowSDN - TFS OSG (https://tfs.etsi.org/) # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json, random, uuid from typing import Dict, Tuple from compute.service.rest_server.nbi_plugins.ietf_network_slice.bindings.network_slice_services import ( NetworkSliceServices ) # R1 emulated devices # Port 13-0 is Optical # Port 13-1 is Copper R1_UUID = "ed2388eb-5fb9-5888-a4f4-160267d3e19b" R1_PORT_13_0_UUID_OPTICAL = "20440915-1a6c-5e7b-a80f-b0e0e51f066d" R1_PORT_13_1_UUID_COPPER = "ff900d5d-2ac0-576c-9628-a2d016681f9d" # R2 emulated devices # Port 13-0 is Optical # Port 13-1 is Copper R2_UUID = "49ce0312-1274-523b-97b8-24d0eca2d72d" R2_PORT_13_0_UUID_OPTICAL = "214618cb-b63b-5e66-84c2-45c1c016e5f0" R2_PORT_13_1_UUID_COPPER = "4e0f7fb4-5d22-56ad-a00e-20bffb4860f9" # R3 emulated devices # Port 13-0 is Optical # Port 13-1 is Copper R3_UUID = "3bc8e994-a3b9-5f60-9c77-6608b1d08313" R3_PORT_13_0_UUID_OPTICAL = "da5196f5-d651-5def-ada6-50ed6430279d" R3_PORT_13_1_UUID_COPPER = "43d221fa-5701-5740-a129-502131f5bda2" # R4 emulated devices # Port 13-0 is Optical # Port 13-1 is Copper R4_UUID = "b43e6361-2573-509d-9a88-1793e751b10d" R4_PORT_13_0_UUID_OPTICAL = "241b74a7-8677-595c-ad65-cc9093c1e341" R4_PORT_13_1_UUID_COPPER = "c57abf46-caaf-5954-90cc-1fec0a69330e" node_dict = {R1_PORT_13_1_UUID_COPPER: R1_UUID, R2_PORT_13_1_UUID_COPPER: R2_UUID, R3_PORT_13_1_UUID_COPPER: R3_UUID, R4_PORT_13_1_UUID_COPPER: R4_UUID} list_endpoints = [R1_PORT_13_1_UUID_COPPER, R2_PORT_13_1_UUID_COPPER, R3_PORT_13_1_UUID_COPPER, R4_PORT_13_1_UUID_COPPER] list_availability= [99, 99.9, 99.99, 99.999, 99.9999] list_bw = [10, 40, 50, 100, 150, 200, 400] list_owner = ["Telefonica", "CTTC", "Telenor", "ADVA", "Ubitech", "ATOS"] URL_POST = "/restconf/data/ietf-network-slice-service:ietf-nss/network-slice-services" URL_DELETE = "/restconf/data/ietf-network-slice-service:ietf-nss/network-slice-services/slice-service=" def generate_request(seed: str) -> Tuple[Dict, str]: ns = NetworkSliceServices() # Slice 1 suuid = str(uuid.uuid5(uuid.NAMESPACE_DNS, str(seed))) slice1 = ns.slice_service[suuid] slice1.service_description = "Test slice for OFC 2023 demo" slice1.status().admin_status().status = "Planned" # TODO not yet mapped # SDPS: R1 optical to R3 optical sdps1 = slice1.sdps().sdp while True: ep1_uuid = random.choice(list_endpoints) ep2_uuid = random.choice(list_endpoints) if ep1_uuid != ep2_uuid: break sdps1[ep1_uuid].node_id = node_dict.get(ep1_uuid) sdps1[ep2_uuid].node_id = node_dict.get(ep2_uuid) # Connectivity group: Connection construct and 2 sla constrains: # - Bandwidth # - Availability cg_uuid = str(uuid.uuid4()) cg = slice1.connection_groups().connection_group cg1 = cg[cg_uuid] cc1 = cg1.connectivity_construct[0] cc1.cc_id = 5 p2p = cc1.connectivity_construct_type.p2p() p2p.p2p_sender_sdp = ep1_uuid p2p.p2p_receiver_sdp = ep2_uuid slo_custom = cc1.slo_sle_policy.custom() metric_bounds = slo_custom.service_slo_sle_policy().metric_bounds().metric_bound # SLO Bandwidth slo_bandwidth = metric_bounds["service-slo-two-way-bandwidth"] slo_bandwidth.value_description = "Guaranteed bandwidth" slo_bandwidth.bound = int(random.choice(list_bw)) slo_bandwidth.metric_unit = "Gbps" # SLO Availability slo_availability = metric_bounds["service-slo-availability"] slo_availability.value_description = "Guaranteed availability" slo_availability.metric_unit = "percentage" slo_availability.bound = random.choice(list_availability) json_request = {"data": ns.to_json()} #Last, add name and owner manually list_name_owner = [{"tag-type": "owner", "value": random.choice(list_owner)}] json_request["data"]["ietf-network-slice-service:network-slice-services"]["slice-service"][0]["service-tags"] = list_name_owner return (json_request, suuid) if __name__ == "__main__": request = generate_request(123) print(json.dumps(request[0], sort_keys=True, indent=4))
src/load_generator/load_gen/RequestGenerator.py +2 −10 Original line number Diff line number Diff line Loading @@ -39,14 +39,6 @@ ROUTER_ID = { 'R149': '5.5.5.5', 'R155': '5.5.5.1', 'R199': '5.5.5.6', } VIRTUAL_CIRCUIT = { 'R149': '5.5.5.5', 'R155': '5.5.5.1', 'R199': '5.5.5.6', } class RequestGenerator: Loading Loading @@ -269,8 +261,8 @@ class RequestGenerator: src_device_name = self._device_data[src_device_uuid]['name'] src_endpoint_name = self._device_endpoint_data[src_device_uuid][src_endpoint_uuid]['name'] src_router_id = ROUTER_ID.get(src_device_name) src_router_num = int(re.findall(r'^\D*(\d+)', src_device_name)[0]) src_router_id = ROUTER_ID.get(src_device_name) if src_router_id is None: src_router_id = '10.0.0.{:d}'.format(src_router_num) dst_device_name = self._device_data[dst_device_uuid]['name'] Loading Loading @@ -322,8 +314,8 @@ class RequestGenerator: src_device_name = self._device_data[src_device_uuid]['name'] src_endpoint_name = self._device_endpoint_data[src_device_uuid][src_endpoint_uuid]['name'] src_router_id = ROUTER_ID.get(src_device_name) src_router_num = int(re.findall(r'^\D*(\d+)', src_device_name)[0]) src_router_id = ROUTER_ID.get(src_device_name) if src_router_id is None: src_router_id = '10.0.0.{:d}'.format(src_router_num) src_address_ip = '10.{:d}.{:d}.{:d}'.format(x, y, src_router_num) Loading
src/pathcomp/frontend/service/algorithms/tools/ComputeSubServices.py +13 −1 Original line number Diff line number Diff line Loading @@ -67,6 +67,7 @@ def convert_explicit_path_hops_to_connections( prv_res_class : Tuple[Optional[int], Optional[DeviceTypeEnum], Optional[str]] = None, None, None for path_hop in path_hops: LOGGER.debug('path_hop={:s}'.format(str(path_hop))) device_uuid = path_hop['device'] if prv_device_uuid == device_uuid: continue device_tuple = device_dict.get(device_uuid) Loading @@ -74,24 +75,33 @@ def convert_explicit_path_hops_to_connections( _,grpc_device = device_tuple res_class = get_resource_classification(grpc_device, device_dict) if res_class[1] in IGNORED_DEVICE_TYPES: continue LOGGER.debug(' prv_res_class={:s}'.format(str(prv_res_class))) LOGGER.debug(' res_class={:s}'.format(str(res_class))) if res_class[1] in IGNORED_DEVICE_TYPES: LOGGER.debug(' ignored') continue if prv_res_class[0] is None: # path ingress LOGGER.debug(' path ingress') connection_stack.put((main_service_uuid, main_service_type, [path_hop], [])) elif prv_res_class[0] > res_class[0]: # create underlying connection LOGGER.debug(' create underlying connection') connection_uuid = str(uuid.uuid4()) prv_service_type = connection_stack.queue[-1][1] service_type = get_service_type(res_class[1], prv_service_type) connection_stack.put((connection_uuid, service_type, [path_hop], [])) elif prv_res_class[0] == res_class[0]: # same resource group kind LOGGER.debug(' same resource group kind') if prv_res_class[1] == res_class[1] and prv_res_class[2] == res_class[2]: # same device type and device controller: connection continues LOGGER.debug(' connection continues') connection_stack.queue[-1][2].append(path_hop) else: # different device type or device controller: chain connections LOGGER.debug(' chain connections') connection = connection_stack.get() connections.append(connection) connection_stack.queue[-1][3].append(connection[0]) Loading @@ -102,6 +112,7 @@ def convert_explicit_path_hops_to_connections( connection_stack.put((connection_uuid, service_type, [path_hop], [])) elif prv_res_class[0] < res_class[0]: # underlying connection ended LOGGER.debug(' underlying connection ended') connection = connection_stack.get() connections.append(connection) connection_stack.queue[-1][3].append(connection[0]) Loading @@ -113,6 +124,7 @@ def convert_explicit_path_hops_to_connections( prv_res_class = res_class # path egress LOGGER.debug(' path egress') connections.append(connection_stack.get()) LOGGER.debug('connections={:s}'.format(str(connections))) assert connection_stack.empty() Loading
src/pathcomp/frontend/service/algorithms/tools/ResourceGroups.py +3 −3 Original line number Diff line number Diff line Loading @@ -33,12 +33,12 @@ DEVICE_TYPE_TO_DEEPNESS = { DeviceTypeEnum.EMULATED_P4_SWITCH.value : 60, DeviceTypeEnum.P4_SWITCH.value : 60, DeviceTypeEnum.EMULATED_MICROWAVE_RADIO_SYSTEM.value : 40, DeviceTypeEnum.MICROWAVE_RADIO_SYSTEM.value : 40, DeviceTypeEnum.EMULATED_XR_CONSTELLATION.value : 40, DeviceTypeEnum.XR_CONSTELLATION.value : 40, DeviceTypeEnum.EMULATED_MICROWAVE_RADIO_SYSTEM.value : 30, DeviceTypeEnum.MICROWAVE_RADIO_SYSTEM.value : 30, DeviceTypeEnum.EMULATED_OPEN_LINE_SYSTEM.value : 30, DeviceTypeEnum.OPEN_LINE_SYSTEM.value : 30, Loading
src/service/service/task_scheduler/TaskExecutor.py +25 −4 Original line number Diff line number Diff line Loading @@ -12,23 +12,28 @@ # See the License for the specific language governing permissions and # limitations under the License. import json import json, logging from enum import Enum from typing import TYPE_CHECKING, Any, Dict, Optional, Union from common.method_wrappers.ServiceExceptions import NotFoundException from common.proto.context_pb2 import Connection, ConnectionId, Device, DeviceId, Service, ServiceId from common.proto.context_pb2 import Connection, ConnectionId, Device, DeviceDriverEnum, DeviceId, Service, ServiceId from common.tools.context_queries.Connection import get_connection_by_id from common.tools.context_queries.Device import get_device from common.tools.context_queries.Service import get_service_by_id from common.tools.grpc.Tools import grpc_message_list_to_json_string from common.tools.object_factory.Device import json_device_id from context.client.ContextClient import ContextClient from device.client.DeviceClient import DeviceClient from service.service.service_handler_api.Exceptions import ( UnsatisfiedFilterException, UnsupportedFilterFieldException, UnsupportedFilterFieldValueException) from service.service.service_handler_api.ServiceHandlerFactory import ServiceHandlerFactory, get_service_handler_class from service.service.tools.ObjectKeys import get_connection_key, get_device_key, get_service_key if TYPE_CHECKING: from service.service.service_handler_api._ServiceHandler import _ServiceHandler LOGGER = logging.getLogger(__name__) CacheableObject = Union[Connection, Device, Service] class CacheableObjectType(Enum): Loading Loading @@ -169,5 +174,21 @@ class TaskExecutor: self, connection : Connection, service : Service, **service_handler_settings ) -> '_ServiceHandler': connection_devices = self.get_devices_from_connection(connection, exclude_managed_by_controller=True) service_handler_class = get_service_handler_class(self._service_handler_factory, service, connection_devices) try: service_handler_class = get_service_handler_class( self._service_handler_factory, service, connection_devices) return service_handler_class(service, self, **service_handler_settings) except (UnsatisfiedFilterException, UnsupportedFilterFieldException, UnsupportedFilterFieldValueException): dict_connection_devices = { cd_data.name : (cd_uuid, cd_data.name, { (device_driver, DeviceDriverEnum.Name(device_driver)) for device_driver in cd_data.device_drivers }) for cd_uuid,cd_data in connection_devices.items() } LOGGER.exception( 'Unable to select service handler. service={:s} connection={:s} connection_devices={:s}'.format( grpc_message_list_to_json_string(service), grpc_message_list_to_json_string(connection), str(dict_connection_devices) ) )