diff options
Diffstat (limited to 'src/vpp-api/python')
-rw-r--r-- | src/vpp-api/python/setup.py | 11 | ||||
-rwxr-xr-x | src/vpp-api/python/tests/test_vpp_serializer.py | 114 | ||||
-rw-r--r-- | src/vpp-api/python/vpp_papi.py (renamed from src/vpp-api/python/vpp_papi/vpp_papi.py) | 393 | ||||
-rw-r--r-- | src/vpp-api/python/vpp_papi/__init__.py | 1 | ||||
-rw-r--r-- | src/vpp-api/python/vpp_papi/vpp_serializer.py | 332 |
5 files changed, 309 insertions, 542 deletions
diff --git a/src/vpp-api/python/setup.py b/src/vpp-api/python/setup.py index 20b9901d521..abda43de606 100644 --- a/src/vpp-api/python/setup.py +++ b/src/vpp-api/python/setup.py @@ -13,20 +13,21 @@ # limitations under the License. try: - from setuptools import setup, find_packages + from setuptools import setup except ImportError: - from distutils.core import setup, find_packages + from distutils.core import setup setup (name = 'vpp_papi', - version = '1.5', + version = '1.4', description = 'VPP Python binding', author = 'Ole Troan', author_email = 'ot@cisco.com', url = 'https://wiki.fd.io/view/VPP/Python_API', + python_requires='>=2.7, >=3.3', license = 'Apache-2.0', test_suite = 'tests', - install_requires=['cffi >= 1.6', 'enum34'], - packages=find_packages(), + install_requires=['cffi >= 1.6'], + py_modules=['vpp_papi'], long_description = '''VPP Python language binding.''', zip_safe = True, ) diff --git a/src/vpp-api/python/tests/test_vpp_serializer.py b/src/vpp-api/python/tests/test_vpp_serializer.py deleted file mode 100755 index 6b867f9e6fe..00000000000 --- a/src/vpp-api/python/tests/test_vpp_serializer.py +++ /dev/null @@ -1,114 +0,0 @@ -#!/usr/bin/env python3 - -import unittest -from vpp_serializer import VPPType, VPPEnumType, VPPUnionType -from socket import inet_pton, AF_INET, AF_INET6 -import logging - -logging.basicConfig(format='%(levelname)s:%(message)s', level=logging.DEBUG) -es_logger = logging.getLogger('vpp_serializer') -es_logger.setLevel(logging.DEBUG) - -class VPPMessage(VPPType): - pass - - -class TestAddType(unittest.TestCase): - - def test_union(self): - un = VPPUnionType('test_union', - [['u8', 'is_bool'], - ['u32', 'is_int']]) - - b = un.pack({'is_int': 0x1234}) - self.assertEqual(len(b), 4) - nt = un.unpack(b) - self.assertEqual(nt.is_bool, 52) - self.assertEqual(nt.is_int, 0x1234) - - def test_address(self): - af = VPPEnumType('vl_api_address_family_t', [["ADDRESS_IP4", 0], - ["ADDRESS_IP6", 1], - {"enumtype": "u32"}]) - ip4 = VPPType('vl_api_ip4_address_t', [['u8', 'address', 4]]) - ip6 = VPPType('vl_api_ip6_address_t', [['u8', 'address', 16]]) - VPPUnionType('vl_api_address_union_t', - [["vl_api_ip4_address_t", "ip4"], - ["vl_api_ip6_address_t", "ip6"]]) - - address = VPPType('address', [['vl_api_address_family_t', 'af'], - ['vl_api_address_union_t', 'un']]) - - b = ip4.pack({'address': inet_pton(AF_INET, '1.1.1.1')}) - self.assertEqual(len(b), 4) - nt = ip4.unpack(b) - self.assertEqual(nt.address, inet_pton(AF_INET, '1.1.1.1')) - - b = ip6.pack({'address': inet_pton(AF_INET6, '1::1')}) - self.assertEqual(len(b), 16) - - b = address.pack({'af': af.ADDRESS_IP4, - 'un': - {'ip4': - {'address': inet_pton(AF_INET, '2.2.2.2')}}}) - self.assertEqual(len(b), 20) - - nt = address.unpack(b) - print('NT union', nt) - self.assertEqual(nt.af, af.ADDRESS_IP4) - self.assertEqual(nt.un.ip4.address, - inet_pton(AF_INET, '2.2.2.2')) - self.assertEqual(nt.un.ip6.address, - inet_pton(AF_INET6, '::0202:0202')) - - def test_arrays(self): - # Test cases - # 1. Fixed list - # 2. Fixed list of variable length sub type - # 3. Variable length type - # - ip4 = VPPType('ip4_address', [['u8', 'address', 4]]) - listip4 = VPPType('list_ip4_t', [['ip4_address', 'addresses', 4]]) - valistip4 = VPPType('list_ip4_t', - [['u8', 'count'], - ['ip4_address', 'addresses', 0, 'count']]) - - valistip4_legacy = VPPType('list_ip4_t', - [['u8', 'foo'], - ['ip4_address', 'addresses', 0]]) - - addresses = [] - for i in range(4): - addresses.append({'address': inet_pton(AF_INET, '2.2.2.2')}) - b = listip4.pack({'addresses': addresses}) - self.assertEqual(len(b), 16) - nt = listip4.unpack(b) - - self.assertEqual(nt.addresses[0].address, - inet_pton(AF_INET, '2.2.2.2')) - - b = valistip4.pack({'count': len(addresses), 'addresses': addresses}) - self.assertEqual(len(b), 17) - - nt = valistip4.unpack(b) - print('NT', nt) - - b = valistip4_legacy.pack({'foo': 1, 'addresses': addresses}) - self.assertEqual(len(b), 17) - nt = valistip4_legacy.unpack(b) - print('NT', nt) - - - def test_message(self): - foo = VPPMessage('foo', [['u16', '_vl_msg_id'], - ['u8', 'client_index'], - ['u8', 'something'], - {"crc": "0x559b9f3c"}]) - b = foo.pack({'_vl_msg_id': 1, 'client_index': 5, - 'something': 200}) - self.assertEqual(len(b), 4) - nt = foo.unpack(b) - print('NT', nt) - -if __name__ == '__main__': - unittest.main() diff --git a/src/vpp-api/python/vpp_papi/vpp_papi.py b/src/vpp-api/python/vpp_papi.py index 5ff8064d425..ece0e4ee52d 100644 --- a/src/vpp-api/python/vpp_papi/vpp_papi.py +++ b/src/vpp-api/python/vpp_papi.py @@ -27,7 +27,6 @@ import weakref import atexit from cffi import FFI import cffi -from vpp_serializer import VPPType, VPPEnumType, VPPUnionType, BaseTypes if sys.version[0] == '2': import Queue as queue @@ -104,9 +103,6 @@ class FuncWrapper(object): return self._func(**kwargs) -class VPPMessage(VPPType): - pass - class VPP(): """VPP interface. @@ -119,56 +115,8 @@ class VPP(): provides a means to register a callback function to receive these messages in a background thread. """ - - def process_json_file(self, apidef_file): - api = json.load(apidef_file) - types = {} - for t in api['enums']: - t[0] = 'vl_api_' + t[0] + '_t' - types[t[0]] = {'type': 'enum', 'data': t} - for t in api['unions']: - t[0] = 'vl_api_' + t[0] + '_t' - types[t[0]] = {'type': 'union', 'data': t} - for t in api['types']: - t[0] = 'vl_api_' + t[0] + '_t' - types[t[0]] = {'type': 'type', 'data': t} - - i = 0 - while True: - unresolved = {} - for k, v in types.items(): - t = v['data'] - if v['type'] == 'enum': - try: - VPPEnumType(t[0], t[1:]) - except ValueError: - unresolved[k] = v - elif v['type'] == 'union': - try: - VPPUnionType(t[0], t[1:]) - except ValueError: - unresolved[k] = v - elif v['type'] == 'type': - try: - VPPType(t[0], t[1:]) - except ValueError: - unresolved[k] = v - if len(unresolved) == 0: - break - if i > 3: - raise ValueError('Unresolved type definitions {}' - .format(unresolved)) - types = unresolved - i += 1 - - for m in api['messages']: - try: - self.messages[m[0]] = VPPMessage(m[0], m[1:]) - except NotImplementedError: - self.logger.error('Not implemented error for {}'.format(m[0])) - def __init__(self, apifiles=None, testmode=False, async_thread=True, - logger=logging.getLogger('vpp_papi'), loglevel='debug', + logger=None, loglevel=None, read_timeout=0): """Create a VPP API object. @@ -189,14 +137,14 @@ class VPP(): logger = logging.getLogger(__name__) if loglevel is not None: logger.setLevel(loglevel) + self.logger = logger self.messages = {} self.id_names = [] self.id_msgdef = [] self.connected = False - self.header = VPPType('header', [['u16', 'msgid'], - ['u32', 'client_index']]) + self.header = struct.Struct('>HI') self.apifiles = [] self.event_callback = None self.message_queue = queue.Queue() @@ -217,8 +165,12 @@ class VPP(): for file in apifiles: with open(file) as apidef_file: - self.process_json_file(apidef_file) + api = json.load(apidef_file) + for t in api['types']: + self.add_type(t[0], t[1:]) + for m in api['messages']: + self.add_message(m[0], m[1:]) self.apifiles = apifiles # Basic sanity check @@ -229,8 +181,7 @@ class VPP(): atexit.register(vpp_atexit, weakref.ref(self)) # Register error handler - if not testmode: - vpp_api.vac_set_error_handler(vac_error_handler) + vpp_api.vac_set_error_handler(vac_error_handler) # Support legacy CFFI # from_buffer supported from 1.8.0 @@ -383,38 +334,305 @@ class VPP(): print('Connected') if self.connected else print('Not Connected') print('Read API definitions from', ', '.join(self.apifiles)) - @property - def api(self): - if not hasattr(self, "_api"): - raise Exception("Not connected, api definitions not available") - return self._api + def __struct(self, t, n=None, e=-1, vl=None): + """Create a packing structure for a message.""" + base_types = {'u8': 'B', + 'u16': 'H', + 'u32': 'I', + 'i32': 'i', + 'u64': 'Q', + 'f64': 'd', } + + if t in base_types: + if not vl: + if e > 0 and t == 'u8': + # Fixed byte array + s = struct.Struct('>' + str(e) + 's') + return s.size, s + if e > 0: + # Fixed array of base type + s = struct.Struct('>' + base_types[t]) + return s.size, [e, s] + elif e == 0: + # Old style variable array + s = struct.Struct('>' + base_types[t]) + return s.size, [-1, s] + else: + # Variable length array + if t == 'u8': + s = struct.Struct('>s') + return s.size, [vl, s] + else: + s = struct.Struct('>' + base_types[t]) + return s.size, [vl, s] + + s = struct.Struct('>' + base_types[t]) + return s.size, s + + if t in self.messages: + size = self.messages[t]['sizes'][0] + + # Return a list in case of array + if e > 0 and not vl: + return size, [e, lambda self, encode, buf, offset, args: ( + self.__struct_type(encode, self.messages[t], buf, offset, + args))] + if vl: + return size, [vl, lambda self, encode, buf, offset, args: ( + self.__struct_type(encode, self.messages[t], buf, offset, + args))] + elif e == 0: + # Old style VLA + raise NotImplementedError(1, + 'No support for compound types ' + t) + return size, lambda self, encode, buf, offset, args: ( + self.__struct_type(encode, self.messages[t], buf, offset, args) + ) + + raise ValueError(1, 'Invalid message type: ' + t) + + def __struct_type(self, encode, msgdef, buf, offset, kwargs): + """Get a message packer or unpacker.""" + if encode: + return self.__struct_type_encode(msgdef, buf, offset, kwargs) + else: + return self.__struct_type_decode(msgdef, buf, offset) + + def __struct_type_encode(self, msgdef, buf, offset, kwargs): + off = offset + size = 0 + + for k in kwargs: + if k not in msgdef['args']: + raise ValueError(1, 'Non existing argument [' + k + ']' + + ' used in call to: ' + + self.id_names[kwargs['_vl_msg_id']] + '()') + + for k, v in vpp_iterator(msgdef['args']): + off += size + if k in kwargs: + if type(v) is list: + if callable(v[1]): + e = kwargs[v[0]] if v[0] in kwargs else v[0] + if e != len(kwargs[k]): + raise (ValueError(1, + 'Input list length mismatch: ' + '%s (%s != %s)' % + (k, e, len(kwargs[k])))) + size = 0 + for i in range(e): + size += v[1](self, True, buf, off + size, + kwargs[k][i]) + else: + if v[0] in kwargs: + kwargslen = kwargs[v[0]] + if kwargslen != len(kwargs[k]): + raise ValueError(1, + 'Input list length mismatch:' + ' %s (%s != %s)' % + (k, kwargslen, + len(kwargs[k]))) + else: + kwargslen = len(kwargs[k]) + if v[1].size == 1: + buf[off:off + kwargslen] = bytearray(kwargs[k]) + size = kwargslen + else: + size = 0 + for i in kwargs[k]: + v[1].pack_into(buf, off + size, i) + size += v[1].size + else: + if callable(v): + size = v(self, True, buf, off, kwargs[k]) + else: + if type(kwargs[k]) is str and v.size < len(kwargs[k]): + raise ValueError(1, + 'Input list length mismatch: ' + '%s (%s < %s)' % + (k, v.size, len(kwargs[k]))) + v.pack_into(buf, off, kwargs[k]) + size = v.size + else: + size = v.size if not type(v) is list else 0 + + return off + size - offset + + def __getitem__(self, name): + if name in self.messages: + return self.messages[name] + return None - def make_function(self, msg, i, multipart, async): + def get_size(self, sizes, kwargs): + total_size = sizes[0] + for e in sizes[1]: + if e in kwargs and type(kwargs[e]) is list: + total_size += len(kwargs[e]) * sizes[1][e] + return total_size + + def encode(self, msgdef, kwargs): + # Make suitably large buffer + size = self.get_size(msgdef['sizes'], kwargs) + buf = bytearray(size) + offset = 0 + size = self.__struct_type(True, msgdef, buf, offset, kwargs) + return buf[:offset + size] + + def decode(self, msgdef, buf): + return self.__struct_type(False, msgdef, buf, 0, None)[1] + + def __struct_type_decode(self, msgdef, buf, offset): + res = [] + off = offset + size = 0 + for k, v in vpp_iterator(msgdef['args']): + off += size + if type(v) is list: + lst = [] + if callable(v[1]): # compound type + size = 0 + if v[0] in msgdef['args']: # vla + e = res[v[2]] + else: # fixed array + e = v[0] + res.append(lst) + for i in range(e): + (s, l) = v[1](self, False, buf, off + size, None) + lst.append(l) + size += s + continue + if v[1].size == 1: + if type(v[0]) is int: + size = len(buf) - off + else: + size = res[v[2]] + res.append(buf[off:off + size]) + else: + e = v[0] if type(v[0]) is int else res[v[2]] + if e == -1: + e = (len(buf) - off) / v[1].size + lst = [] + res.append(lst) + size = 0 + for i in range(e): + lst.append(v[1].unpack_from(buf, off + size)[0]) + size += v[1].size + else: + if callable(v): + size = 0 + (s, l) = v(self, False, buf, off, None) + res.append(l) + size += s + else: + res.append(v.unpack_from(buf, off)[0]) + size = v.size + + return off + size - offset, msgdef['return_tuple']._make(res) + + def ret_tup(self, name): + if name in self.messages and 'return_tuple' in self.messages[name]: + return self.messages[name]['return_tuple'] + return None + + def duplicate_check_ok(self, name, msgdef): + crc = None + for c in msgdef: + if type(c) is dict and 'crc' in c: + crc = c['crc'] + break + if crc: + # We can get duplicates because of imports + if crc == self.messages[name]['crc']: + return True + return False + + def add_message(self, name, msgdef, typeonly=False): + if name in self.messages: + if typeonly: + if self.duplicate_check_ok(name, msgdef): + return + raise ValueError('Duplicate message name: ' + name) + + args = collections.OrderedDict() + argtypes = collections.OrderedDict() + fields = [] + msg = {} + total_size = 0 + sizes = {} + for i, f in enumerate(msgdef): + if type(f) is dict and 'crc' in f: + msg['crc'] = f['crc'] + continue + field_type = f[0] + field_name = f[1] + if len(f) == 3 and f[2] == 0 and i != len(msgdef) - 2: + raise ValueError('Variable Length Array must be last: ' + name) + size, s = self.__struct(*f) + args[field_name] = s + if type(s) == list and type(s[0]) == int and \ + type(s[1]) == struct.Struct: + if s[0] < 0: + sizes[field_name] = size + else: + sizes[field_name] = size + total_size += s[0] * size + else: + sizes[field_name] = size + total_size += size + + argtypes[field_name] = field_type + if len(f) == 4: # Find offset to # elements field + idx = list(args.keys()).index(f[3]) - i + args[field_name].append(idx) + fields.append(field_name) + msg['return_tuple'] = collections.namedtuple(name, fields, + rename=True) + self.messages[name] = msg + self.messages[name]['args'] = args + self.messages[name]['argtypes'] = argtypes + self.messages[name]['typeonly'] = typeonly + self.messages[name]['sizes'] = [total_size, sizes] + return self.messages[name] + + def add_type(self, name, typedef): + return self.add_message('vl_api_' + name + '_t', typedef, + typeonly=True) + + def make_function(self, name, i, msgdef, multipart, async): if (async): def f(**kwargs): - return self._call_vpp_async(i, msg, **kwargs) + return self._call_vpp_async(i, msgdef, **kwargs) else: def f(**kwargs): - return self._call_vpp(i, msg, multipart, **kwargs) - - f.__name__ = str(msg.name) + return self._call_vpp(i, msgdef, multipart, **kwargs) + args = self.messages[name]['args'] + argtypes = self.messages[name]['argtypes'] + f.__name__ = str(name) f.__doc__ = ", ".join(["%s %s" % - (msg.fieldtypes[j], k) for j, k in enumerate(msg.fields)]) + (argtypes[k], k) for k in args.keys()]) return f + @property + def api(self): + if not hasattr(self, "_api"): + raise Exception("Not connected, api definitions not available") + return self._api + def _register_functions(self, async=False): self.id_names = [None] * (self.vpp_dictionary_maxid + 1) self.id_msgdef = [None] * (self.vpp_dictionary_maxid + 1) self._api = Empty() - for name, msg in vpp_iterator(self.messages): - n = name + '_' + msg.crc[2:] + for name, msgdef in vpp_iterator(self.messages): + if self.messages[name]['typeonly']: + continue + crc = self.messages[name]['crc'] + n = name + '_' + crc[2:] i = vpp_api.vac_get_msg_index(n.encode()) if i > 0: - self.id_msgdef[i] = msg + self.id_msgdef[i] = msgdef self.id_names[i] = name - # TODO: Fix multipart (use services) multipart = True if name.find('_dump') > 0 else False - f = self.make_function(msg, i, multipart, async) + f = self.make_function(name, i, msgdef, multipart, async) setattr(self._api, name, FuncWrapper(f)) else: self.logger.debug( @@ -451,11 +669,12 @@ class VPP(): if rv != 0: raise IOError(2, 'Connect failed') self.connected = True + self.vpp_dictionary_maxid = vpp_api.vac_msg_table_max_index() self._register_functions(async=async) # Initialise control ping - crc = self.messages['control_ping'].crc + crc = self.messages['control_ping']['crc'] self.control_ping_index = vpp_api.vac_get_msg_index( ('control_ping' + '_' + crc[2:]).encode()) self.control_ping_msgdef = self.messages['control_ping'] @@ -524,18 +743,18 @@ class VPP(): self.logger.warning('vpp_api.read failed') return - i, ci = self.header.unpack(msg, 0) + i, ci = self.header.unpack_from(msg, 0) if self.id_names[i] == 'rx_thread_exit': return # # Decode message and returns a tuple. # - msgobj = self.id_msgdef[i] - if not msgobj: + msgdef = self.id_msgdef[i] + if not msgdef: raise IOError(2, 'Reply message undefined') - r = msgobj.unpack(msg) + r = self.decode(msgdef, msg) return r @@ -559,12 +778,7 @@ class VPP(): self.control_ping_msgdef, context=context) - def validate_args(self, msg, kwargs): - d = set(kwargs.keys()) - set(msg.field_by_name.keys()) - if d: - raise ValueError('Invalid argument {} to {}'.format(list(d), msg.name)) - - def _call_vpp(self, i, msg, multipart, **kwargs): + def _call_vpp(self, i, msgdef, multipart, **kwargs): """Given a message, send the message and await a reply. msgdef - the message packing definition @@ -586,9 +800,8 @@ class VPP(): else: context = kwargs['context'] kwargs['_vl_msg_id'] = i + b = self.encode(msgdef, kwargs) - self.validate_args(msg, kwargs) - b = msg.pack(kwargs) vpp_api.vac_rx_suspend() self._write(b) @@ -603,6 +816,7 @@ class VPP(): msg = self._read() if not msg: raise IOError(2, 'VPP API client: read failed') + r = self.decode_incoming_msg(msg) msgname = type(r).__name__ if context not in r or r.context == 0 or context != r.context: @@ -621,7 +835,7 @@ class VPP(): return rl - def _call_vpp_async(self, i, msg, **kwargs): + def _call_vpp_async(self, i, msgdef, **kwargs): """Given a message, send the message and await a reply. msgdef - the message packing definition @@ -635,9 +849,8 @@ class VPP(): kwargs['context'] = context else: context = kwargs['context'] - kwargs['client_index'] = 0 kwargs['_vl_msg_id'] = i - b = msg.pack(kwargs) + b = self.encode(msgdef, kwargs) self._write(b) diff --git a/src/vpp-api/python/vpp_papi/__init__.py b/src/vpp-api/python/vpp_papi/__init__.py deleted file mode 100644 index f9afcf17f29..00000000000 --- a/src/vpp-api/python/vpp_papi/__init__.py +++ /dev/null @@ -1 +0,0 @@ -from .vpp_papi import * diff --git a/src/vpp-api/python/vpp_papi/vpp_serializer.py b/src/vpp-api/python/vpp_papi/vpp_serializer.py deleted file mode 100644 index 146a8f6919a..00000000000 --- a/src/vpp-api/python/vpp_papi/vpp_serializer.py +++ /dev/null @@ -1,332 +0,0 @@ -# -# Copyright (c) 2018 Cisco and/or its affiliates. -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at: -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# - -import struct -import collections -from enum import IntEnum -import logging - -# -# Set log-level in application by doing e.g.: -# logger = logging.getLogger('vpp_serializer') -# logger.setLevel(logging.DEBUG) -# -logger = logging.getLogger(__name__) -FORMAT = "[%(filename)s:%(lineno)s - %(funcName)s() ] %(message)s" -logging.basicConfig(format=FORMAT) - - -class BaseTypes(): - def __init__(self, type, elements=0): - base_types = {'u8': '>B', - 'u16': '>H', - 'u32': '>I', - 'i32': '>i', - 'u64': '>Q', - 'f64': '>d', - 'header': '>HI'} - - if elements > 0 and type == 'u8': - self.packer = struct.Struct('>%ss' % elements) - else: - self.packer = struct.Struct(base_types[type]) - self.size = self.packer.size - logger.debug('Adding {} with format: {}' - .format(type, base_types[type])) - - def pack(self, data, kwargs=None): - logger.debug("Data: {} Format: {}".format(data, self.packer.format)) - return self.packer.pack(data) - - def unpack(self, data, offset, result=None): - logger.debug("@ {} Format: {}".format(offset, self.packer.format)) - return self.packer.unpack_from(data, offset)[0] - - -types = {} -types['u8'] = BaseTypes('u8') -types['u16'] = BaseTypes('u16') -types['u32'] = BaseTypes('u32') -types['i32'] = BaseTypes('i32') -types['u64'] = BaseTypes('u64') -types['f64'] = BaseTypes('f64') - - -class FixedList_u8(): - def __init__(self, name, field_type, num): - self.name = name - self.num = num - self.packer = BaseTypes(field_type, num) - self.size = self.packer.size - - def pack(self, list, kwargs): - logger.debug("Data: {}".format(list)) - - if len(list) > self.num: - raise ValueError('Fixed list length error for "{}", got: {}' - ' expected: {}' - .format(self.name, len(list), self.num)) - return self.packer.pack(list) - - def unpack(self, data, offset=0, result=None): - if len(data[offset:]) < self.num: - raise ValueError('Invalid array length for "{}" got {}' - ' expected {}' - .format(self.name, len(data), self.num)) - return self.packer.unpack(data, offset) - - -class FixedList(): - def __init__(self, name, field_type, num): - self.num = num - self.packer = types[field_type] - self.size = self.packer.size * num - - def pack(self, list, kwargs): - logger.debug("Data: {}".format(list)) - - if len(list) != self.num: - raise ValueError('Fixed list length error, got: {} expected: {}' - .format(len(list), self.num)) - b = bytes() - for e in list: - b += self.packer.pack(e) - return b - - def unpack(self, data, offset=0, result=None): - # Return a list of arguments - result = [] - for e in range(self.num): - x = self.packer.unpack(data, offset) - result.append(x) - offset += self.packer.size - return result - - -class VLAList(): - def __init__(self, name, field_type, len_field_name, index): - self.index = index - self.packer = types[field_type] - self.size = self.packer.size - self.length_field = len_field_name - - def pack(self, list, kwargs=None): - logger.debug("Data: {}".format(list)) - if len(list) != kwargs[self.length_field]: - raise ValueError('Variable length error, got: {} expected: {}' - .format(len(list), kwargs[self.length_field])) - b = bytes() - - # u8 array - if self.packer.size == 1: - p = BaseTypes('u8', len(list)) - return p.pack(list) - - for e in list: - b += self.packer.pack(e) - return b - - def unpack(self, data, offset=0, result=None): - logger.debug("Data: {} @ {} Result: {}" - .format(list, offset, result[self.index])) - # Return a list of arguments - - # u8 array - if self.packer.size == 1: - if result[self.index] == 0: - return b'' - p = BaseTypes('u8', result[self.index]) - r = p.unpack(data, offset) - return r - - r = [] - for e in range(result[self.index]): - x = self.packer.unpack(data, offset) - r.append(x) - offset += self.packer.size - return r - - -class VLAList_legacy(): - def __init__(self, name, field_type): - self.packer = types[field_type] - self.size = self.packer.size - - def pack(self, list, kwargs=None): - logger.debug("Data: {}".format(list)) - b = bytes() - for e in list: - b += self.packer.pack(e) - return b - - def unpack(self, data, offset=0, result=None): - # Return a list of arguments - if (len(data) - offset) % self.packer.size: - raise ValueError('Legacy Variable Length Array length mismatch.') - elements = int((len(data) - offset) / self.packer.size) - r = [] - logger.debug("Legacy VLA: {} elements of size {}" - .format(elements, self.packer.size)) - for e in range(elements): - x = self.packer.unpack(data, offset) - r.append(x) - offset += self.packer.size - return r - - -class VPPEnumType(): - def __init__(self, name, msgdef): - self.size = types['u32'].size - e_hash = {} - for f in msgdef: - if type(f) is dict and 'enumtype' in f: - if f['enumtype'] != 'u32': - raise NotImplementedError - continue - ename, evalue = f - e_hash[ename] = evalue - self.enum = IntEnum(name, e_hash) - types[name] = self - logger.debug('Adding enum {}'.format(name)) - - def __getattr__(self, name): - return self.enum[name] - - def pack(self, data, kwargs=None): - logger.debug("Data: {}".format(data)) - return types['u32'].pack(data, kwargs) - - def unpack(self, data, offset=0, result=None): - x = types['u32'].unpack(data, offset) - return self.enum(x) - - -class VPPUnionType(): - def __init__(self, name, msgdef): - self.name = name - self.size = 0 - self.maxindex = 0 - fields = [] - self.packers = collections.OrderedDict() - for i, f in enumerate(msgdef): - if type(f) is dict and 'crc' in f: - self.crc = f['crc'] - continue - f_type, f_name = f - if f_type not in types: - logger.debug('Unknown union type {}'.format(f_type)) - raise ValueError('Unknown message type {}'.format(f_type)) - fields.append(f_name) - size = types[f_type].size - self.packers[f_name] = types[f_type] - if size > self.size: - self.size = size - self.maxindex = i - - types[name] = self - self.tuple = collections.namedtuple(name, fields, rename=True) - logger.debug('Adding union {}'.format(name)) - - def pack(self, data, kwargs=None): - logger.debug("Data: {}".format(data)) - for k, v in data.items(): - logger.debug("Key: {} Value: {}".format(k, v)) - b = self.packers[k].pack(v, kwargs) - offset = self.size - self.packers[k].size - break - r = bytearray(self.size) - r[offset:] = b - return r - - def unpack(self, data, offset=0, result=None): - r = [] - for k, p in self.packers.items(): - union_offset = self.size - p.size - r.append(p.unpack(data, offset + union_offset)) - return self.tuple._make(r) - - -class VPPType(): - # Set everything up to be able to pack / unpack - def __init__(self, name, msgdef): - self.name = name - self.msgdef = msgdef - self.packers = [] - self.fields = [] - self.fieldtypes = [] - self.field_by_name = {} - size = 0 - for i, f in enumerate(msgdef): - if type(f) is dict and 'crc' in f: - self.crc = f['crc'] - continue - f_type, f_name = f[:2] - self.fields.append(f_name) - self.field_by_name[f_name] = None - self.fieldtypes.append(f_type) - if f_type not in types: - logger.debug('Unknown type {}'.format(f_type)) - raise ValueError('Unknown message type {}'.format(f_type)) - if len(f) == 3: # list - list_elements = f[2] - if list_elements == 0: - p = VLAList_legacy(f_name, f_type) - self.packers.append(p) - elif f_type == 'u8': - p = FixedList_u8(f_name, f_type, list_elements) - self.packers.append(p) - size += p.size - else: - p = FixedList(f_name, f_type, list_elements) - self.packers.append(p) - size += p.size - elif len(f) == 4: # Variable length list - # Find index of length field - length_index = self.fields.index(f[3]) - p = VLAList(f_name, f_type, f[3], length_index) - self.packers.append(p) - else: - self.packers.append(types[f_type]) - size += types[f_type].size - - self.size = size - self.tuple = collections.namedtuple(name, self.fields, rename=True) - types[name] = self - logger.debug('Adding type {}'.format(name)) - - def pack(self, data, kwargs=None): - if not kwargs: - kwargs = data - logger.debug("Data: {}".format(data)) - b = bytes() - for i, a in enumerate(self.fields): - if a not in data: - logger.debug("Argument {} not given, defaulting to 0" - .format(a)) - b += b'\x00' * self.packers[i].size - continue - b += self.packers[i].pack(data[a], kwargs) - return b - - def unpack(self, data, offset=0, result=None): - # Return a list of arguments - result = [] - for p in self.packers: - x = p.unpack(data, offset, result) - if type(x) is tuple and len(x) == 1: - x = x[0] - result.append(x) - offset += p.size - return self.tuple._make(result) |