# Copyright 2021-2023 H2020 TeraFlow (https://www.teraflow-h2020.eu/)
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import grpc, json, logging, requests, uuid
from typing import Dict, Tuple
from common.proto.context_pb2 import Empty, EndPointId, Service
from common.proto.pathcomp_pb2 import PathCompReply, PathCompRequest
from common.proto.pathcomp_pb2_grpc import PathCompServiceServicer
from common.rpc_method_wrapper.Decorator import create_metrics, safe_and_metered_rpc_method
from common.tools.grpc.Tools import grpc_message_to_json_string
from context.client.ContextClient import ContextClient
from pathcomp.frontend.Config import BACKEND_URL
from pathcomp.frontend.service.tools.ComposeRequest import compose_device, compose_link, compose_service
#from pathcomp.frontend.service.tools.Constants import CapacityUnit
LOGGER = logging.getLogger(__name__)
SERVICE_NAME = 'PathComp'
METHOD_NAMES = ['Compute']
METRICS = create_metrics(SERVICE_NAME, METHOD_NAMES)
class PathCompServiceServicerImpl(PathCompServiceServicer):
def __init__(self) -> None:
LOGGER.debug('Creating Servicer...')
LOGGER.debug('Servicer Created')
@safe_and_metered_rpc_method(METRICS, LOGGER)
def Compute(self, request : PathCompRequest, context : grpc.ServicerContext) -> PathCompReply:
LOGGER.info('[Compute] begin ; request = {:s}'.format(grpc_message_to_json_string(request)))
algorithm = request.WhichOneof('algorithm')
if algorithm == 'shortest_path':
# no attributes
pass
elif algorithm == 'k_shortest_path':
k_inspection = request.k_shortest_path.k_inspection
k_return = request.k_shortest_path.k_return
else:
raise NotImplementedError('Unsupported Algorithm: {:s}'.format(str(algorithm)))
context_client = ContextClient()
algorithm = {'id': 'KSP', 'sync': False, 'k_paths': k_return}
service_list = [
compose_service(grpc_service, algorithm)
for grpc_service in request.services
]
get_service_key = lambda service_id: (service_id['contextId'], service_id['service_uuid'])
service_dict : Dict[Tuple[str, str], Tuple[Dict, Service]] = {
get_service_key(json_service['serviceId']): (json_service, grpc_service)
for json_service,grpc_service in zip(service_list, request.services)
}
#LOGGER.info('service_dict = {:s}'.format(str(service_dict)))
# TODO: consider filtering resources
#grpc_contexts = context_client.ListContexts(Empty())
#for grpc_context in grpc_contexts.contexts:
# # TODO: add context to request
# grpc_topologies = context_client.ListTopologies(grpc_context.context_id)
# for grpc_topology in grpc_topologies.topologies: #pylint: disable=unused-variable
# # TODO: add topology to request
# pass
grpc_devices = context_client.ListDevices(Empty())
device_list = [
compose_device(grpc_device)
for grpc_device in grpc_devices.devices
]
endpoint_dict : Dict[str, Dict[str, Tuple[Dict, EndPointId]]] = {
json_device['device_Id']: {
json_endpoint['endpoint_id']['endpoint_uuid']: (json_endpoint['endpoint_id'], grpc_endpoint.endpoint_id)
for json_endpoint,grpc_endpoint in zip(json_device['device_endpoints'], grpc_device.device_endpoints)
}
for json_device,grpc_device in zip(device_list, grpc_devices.devices)
}
#LOGGER.info('endpoint_dict = {:s}'.format(str(endpoint_dict)))
grpc_links = context_client.ListLinks(Empty())
link_list = [
compose_link(grpc_link)
for grpc_link in grpc_links.links
]
request = {
'serviceList': service_list,
'deviceList' : device_list,
'linkList' : link_list,
}
#with open('pc-req.json', 'w', encoding='UTF-8') as f:
# f.write(json.dumps(request, sort_keys=True, indent=4))
reply = requests.post(BACKEND_URL, json=request)
if reply.status_code not in {requests.codes.ok}:
raise Exception('Backend error({:s}) for request({:s})'.format(
str(reply.content.decode('UTF-8')), json.dumps(request, sort_keys=True)))
LOGGER.info('status_code={:s} reply={:s}'.format(
str(reply.status_code), str(reply.content.decode('UTF-8'))))
json_reply = reply.json()
response_list = json_reply.get('response-list', [])
reply = PathCompReply()
for response in response_list:
service_key = get_service_key(response['serviceId'])
tuple_service = service_dict.get(service_key)
if tuple_service is None: raise Exception('ServiceKey({:s}) not found'.format(str(service_key)))
json_service, grpc_service = tuple_service
# TODO: implement support for multi-point services
service_endpoint_ids = grpc_service.service_endpoint_ids
if len(service_endpoint_ids) != 2: raise NotImplementedError('Service must have 2 endpoints')
service = reply.services.add()
service.CopyFrom(grpc_service)
connection = reply.connections.add()
connection.connection_id.connection_uuid.uuid = str(uuid.uuid4())
connection.service_id.CopyFrom(service.service_id)
no_path_issue = response.get('noPath', {}).get('issue')
if no_path_issue is not None:
# no path found: leave connection with no endpoints
# no_path_issue == 1 => no path due to a constraint
continue
service_paths = response['path']
for service_path in service_paths:
# ... "path-capacity": {"total-size": {"value": 200, "unit": 0}},
# ... "path-latency": {"fixed-latency-characteristic": "10.000000"},
# ... "path-cost": {"cost-name": "", "cost-value": "5.000000", "cost-algorithm": "0.000000"},
#path_capacity = service_path['path-capacity']['total-size']
#path_capacity_value = path_capacity['value']
#path_capacity_unit = CapacityUnit(path_capacity['unit'])
#path_latency = service_path['path-latency']['fixed-latency-characteristic']
#path_cost = service_path['path-cost']
#path_cost_name = path_cost['cost-name']
#path_cost_value = path_cost['cost-value']
#path_cost_algorithm = path_cost['cost-algorithm']
path_endpoints = service_path['devices']
for endpoint in path_endpoints:
device_uuid = endpoint['device_id']
endpoint_uuid = endpoint['endpoint_uuid']
endpoint_id = connection.path_hops_endpoint_ids.add()
endpoint_id.CopyFrom(endpoint_dict[device_uuid][endpoint_uuid][1])
LOGGER.info('[Compute] end ; reply = {:s}'.format(grpc_message_to_json_string(reply)))
return reply