diff options
author | Ole Trøan <otroan@employees.org> | 2018-06-18 18:30:09 +0000 |
---|---|---|
committer | Ole Trøan <otroan@employees.org> | 2018-06-18 18:30:09 +0000 |
commit | 61debaaaf2230c75d2435a6c866293e2d6359c44 (patch) | |
tree | f39dcd6ddc1ba13eeb08d29de8530dde8999ce2a /src/vpp-api/python/vpp_papi | |
parent | a5ee900fb75201bbfceaf13c8bc57a13ed094988 (diff) |
Revert "Python API: Add enum and union support."
This reverts commit a5ee900fb75201bbfceaf13c8bc57a13ed094988.
Some of the unit tests breaks. Backing out until fixed.
Change-Id: I1846fb417db44a2a772f7b59cda8bcfe6d39f8c3
Signed-off-by: Ole Troan <ot@cisco.com>
Diffstat (limited to 'src/vpp-api/python/vpp_papi')
-rw-r--r-- | src/vpp-api/python/vpp_papi/__init__.py | 1 | ||||
-rw-r--r-- | src/vpp-api/python/vpp_papi/vpp_papi.py | 678 | ||||
-rw-r--r-- | src/vpp-api/python/vpp_papi/vpp_serializer.py | 332 |
3 files changed, 0 insertions, 1011 deletions
diff --git a/src/vpp-api/python/vpp_papi/__init__.py b/src/vpp-api/python/vpp_papi/__init__.py deleted file mode 100644 index f9afcf17f29..00000000000 --- a/src/vpp-api/python/vpp_papi/__init__.py +++ /dev/null @@ -1 +0,0 @@ -from .vpp_papi import * diff --git a/src/vpp-api/python/vpp_papi/vpp_papi.py b/src/vpp-api/python/vpp_papi/vpp_papi.py deleted file mode 100644 index 5ff8064d425..00000000000 --- a/src/vpp-api/python/vpp_papi/vpp_papi.py +++ /dev/null @@ -1,678 +0,0 @@ -#!/usr/bin/env python -# -# Copyright (c) 2016 Cisco and/or its affiliates. -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at: -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# - -from __future__ import print_function -import sys -import os -import logging -import collections -import struct -import json -import threading -import fnmatch -import weakref -import atexit -from cffi import FFI -import cffi -from vpp_serializer import VPPType, VPPEnumType, VPPUnionType, BaseTypes - -if sys.version[0] == '2': - import Queue as queue -else: - import queue as queue - -ffi = FFI() -ffi.cdef(""" -typedef void (*vac_callback_t)(unsigned char * data, int len); -typedef void (*vac_error_callback_t)(void *, unsigned char *, int); -int vac_connect(char * name, char * chroot_prefix, vac_callback_t cb, - int rx_qlen); -int vac_disconnect(void); -int vac_read(char **data, int *l, unsigned short timeout); -int vac_write(char *data, int len); -void vac_free(void * msg); - -int vac_get_msg_index(unsigned char * name); -int vac_msg_table_size(void); -int vac_msg_table_max_index(void); - -void vac_rx_suspend (void); -void vac_rx_resume (void); -void vac_set_error_handler(vac_error_callback_t); - """) - -# Barfs on failure, no need to check success. -vpp_api = ffi.dlopen('libvppapiclient.so') - - -def vpp_atexit(vpp_weakref): - """Clean up VPP connection on shutdown.""" - vpp_instance = vpp_weakref() - if vpp_instance.connected: - vpp_instance.logger.debug('Cleaning up VPP on exit') - vpp_instance.disconnect() - - -vpp_object = None - - -def vpp_iterator(d): - if sys.version[0] == '2': - return d.iteritems() - else: - return d.items() - - -@ffi.callback("void(unsigned char *, int)") -def vac_callback_sync(data, len): - vpp_object.msg_handler_sync(ffi.buffer(data, len)) - - -@ffi.callback("void(unsigned char *, int)") -def vac_callback_async(data, len): - vpp_object.msg_handler_async(ffi.buffer(data, len)) - - -@ffi.callback("void(void *, unsigned char *, int)") -def vac_error_handler(arg, msg, msg_len): - vpp_object.logger.warning("VPP API client:: %s", ffi.string(msg, msg_len)) - - -class Empty(object): - pass - - -class FuncWrapper(object): - def __init__(self, func): - self._func = func - self.__name__ = func.__name__ - - def __call__(self, **kwargs): - return self._func(**kwargs) - - -class VPPMessage(VPPType): - pass - -class VPP(): - """VPP interface. - - This class provides the APIs to VPP. The APIs are loaded - from provided .api.json files and makes functions accordingly. - These functions are documented in the VPP .api files, as they - are dynamically created. - - Additionally, VPP can send callback messages; this class - provides a means to register a callback function to receive - these messages in a background thread. - """ - - def process_json_file(self, apidef_file): - api = json.load(apidef_file) - types = {} - for t in api['enums']: - t[0] = 'vl_api_' + t[0] + '_t' - types[t[0]] = {'type': 'enum', 'data': t} - for t in api['unions']: - t[0] = 'vl_api_' + t[0] + '_t' - types[t[0]] = {'type': 'union', 'data': t} - for t in api['types']: - t[0] = 'vl_api_' + t[0] + '_t' - types[t[0]] = {'type': 'type', 'data': t} - - i = 0 - while True: - unresolved = {} - for k, v in types.items(): - t = v['data'] - if v['type'] == 'enum': - try: - VPPEnumType(t[0], t[1:]) - except ValueError: - unresolved[k] = v - elif v['type'] == 'union': - try: - VPPUnionType(t[0], t[1:]) - except ValueError: - unresolved[k] = v - elif v['type'] == 'type': - try: - VPPType(t[0], t[1:]) - except ValueError: - unresolved[k] = v - if len(unresolved) == 0: - break - if i > 3: - raise ValueError('Unresolved type definitions {}' - .format(unresolved)) - types = unresolved - i += 1 - - for m in api['messages']: - try: - self.messages[m[0]] = VPPMessage(m[0], m[1:]) - except NotImplementedError: - self.logger.error('Not implemented error for {}'.format(m[0])) - - def __init__(self, apifiles=None, testmode=False, async_thread=True, - logger=logging.getLogger('vpp_papi'), loglevel='debug', - read_timeout=0): - """Create a VPP API object. - - apifiles is a list of files containing API - descriptions that will be loaded - methods will be - dynamically created reflecting these APIs. If not - provided this will load the API files from VPP's - default install location. - - logger, if supplied, is the logging logger object to log to. - loglevel, if supplied, is the log level this logger is set - to report at (from the loglevels in the logging module). - """ - global vpp_object - vpp_object = self - - if logger is None: - logger = logging.getLogger(__name__) - if loglevel is not None: - logger.setLevel(loglevel) - self.logger = logger - - self.messages = {} - self.id_names = [] - self.id_msgdef = [] - self.connected = False - self.header = VPPType('header', [['u16', 'msgid'], - ['u32', 'client_index']]) - self.apifiles = [] - self.event_callback = None - self.message_queue = queue.Queue() - self.read_timeout = read_timeout - self.vpp_api = vpp_api - self.async_thread = async_thread - - if not apifiles: - # Pick up API definitions from default directory - try: - apifiles = self.find_api_files() - except RuntimeError: - # In test mode we don't care that we can't find the API files - if testmode: - apifiles = [] - else: - raise - - for file in apifiles: - with open(file) as apidef_file: - self.process_json_file(apidef_file) - - self.apifiles = apifiles - - # Basic sanity check - if len(self.messages) == 0 and not testmode: - raise ValueError(1, 'Missing JSON message definitions') - - # Make sure we allow VPP to clean up the message rings. - atexit.register(vpp_atexit, weakref.ref(self)) - - # Register error handler - if not testmode: - vpp_api.vac_set_error_handler(vac_error_handler) - - # Support legacy CFFI - # from_buffer supported from 1.8.0 - (major, minor, patch) = [int(s) for s in - cffi.__version__.split('.', 3)] - if major >= 1 and minor >= 8: - self._write = self._write_new_cffi - else: - self._write = self._write_legacy_cffi - - class ContextId(object): - """Thread-safe provider of unique context IDs.""" - def __init__(self): - self.context = 0 - self.lock = threading.Lock() - - def __call__(self): - """Get a new unique (or, at least, not recently used) context.""" - with self.lock: - self.context += 1 - return self.context - get_context = ContextId() - - @classmethod - def find_api_dir(cls): - """Attempt to find the best directory in which API definition - files may reside. If the value VPP_API_DIR exists in the environment - then it is first on the search list. If we're inside a recognized - location in a VPP source tree (src/scripts and src/vpp-api/python) - then entries from there to the likely locations in build-root are - added. Finally the location used by system packages is added. - - :returns: A single directory name, or None if no such directory - could be found. - """ - dirs = [] - - if 'VPP_API_DIR' in os.environ: - dirs.append(os.environ['VPP_API_DIR']) - - # perhaps we're in the 'src/scripts' or 'src/vpp-api/python' dir; - # in which case, plot a course to likely places in the src tree - import __main__ as main - if hasattr(main, '__file__'): - # get the path of the calling script - localdir = os.path.dirname(os.path.realpath(main.__file__)) - else: - # use cwd if there is no calling script - localdir = os.getcwd() - localdir_s = localdir.split(os.path.sep) - - def dmatch(dir): - """Match dir against right-hand components of the script dir""" - d = dir.split('/') # param 'dir' assumes a / separator - length = len(d) - return len(localdir_s) > length and localdir_s[-length:] == d - - def sdir(srcdir, variant): - """Build a path from srcdir to the staged API files of - 'variant' (typically '' or '_debug')""" - # Since 'core' and 'plugin' files are staged - # in separate directories, we target the parent dir. - return os.path.sep.join(( - srcdir, - 'build-root', - 'install-vpp%s-native' % variant, - 'vpp', - 'share', - 'vpp', - 'api', - )) - - srcdir = None - if dmatch('src/scripts'): - srcdir = os.path.sep.join(localdir_s[:-2]) - elif dmatch('src/vpp-api/python'): - srcdir = os.path.sep.join(localdir_s[:-3]) - elif dmatch('test'): - # we're apparently running tests - srcdir = os.path.sep.join(localdir_s[:-1]) - - if srcdir: - # we're in the source tree, try both the debug and release - # variants. - dirs.append(sdir(srcdir, '_debug')) - dirs.append(sdir(srcdir, '')) - - # Test for staged copies of the scripts - # For these, since we explicitly know if we're running a debug versus - # release variant, target only the relevant directory - if dmatch('build-root/install-vpp_debug-native/vpp/bin'): - srcdir = os.path.sep.join(localdir_s[:-4]) - dirs.append(sdir(srcdir, '_debug')) - if dmatch('build-root/install-vpp-native/vpp/bin'): - srcdir = os.path.sep.join(localdir_s[:-4]) - dirs.append(sdir(srcdir, '')) - - # finally, try the location system packages typically install into - dirs.append(os.path.sep.join(('', 'usr', 'share', 'vpp', 'api'))) - - # check the directories for existance; first one wins - for dir in dirs: - if os.path.isdir(dir): - return dir - - return None - - @classmethod - def find_api_files(cls, api_dir=None, patterns='*'): - """Find API definition files from the given directory tree with the - given pattern. If no directory is given then find_api_dir() is used - to locate one. If no pattern is given then all definition files found - in the directory tree are used. - - :param api_dir: A directory tree in which to locate API definition - files; subdirectories are descended into. - If this is None then find_api_dir() is called to discover it. - :param patterns: A list of patterns to use in each visited directory - when looking for files. - This can be a list/tuple object or a comma-separated string of - patterns. Each value in the list will have leading/trialing - whitespace stripped. - The pattern specifies the first part of the filename, '.api.json' - is appended. - The results are de-duplicated, thus overlapping patterns are fine. - If this is None it defaults to '*' meaning "all API files". - :returns: A list of file paths for the API files found. - """ - if api_dir is None: - api_dir = cls.find_api_dir() - if api_dir is None: - raise RuntimeError("api_dir cannot be located") - - if isinstance(patterns, list) or isinstance(patterns, tuple): - patterns = [p.strip() + '.api.json' for p in patterns] - else: - patterns = [p.strip() + '.api.json' for p in patterns.split(",")] - - api_files = [] - for root, dirnames, files in os.walk(api_dir): - # iterate all given patterns and de-dup the result - files = set(sum([fnmatch.filter(files, p) for p in patterns], [])) - for filename in files: - api_files.append(os.path.join(root, filename)) - - return api_files - - def status(self): - """Debug function: report current VPP API status to stdout.""" - print('Connected') if self.connected else print('Not Connected') - print('Read API definitions from', ', '.join(self.apifiles)) - - @property - def api(self): - if not hasattr(self, "_api"): - raise Exception("Not connected, api definitions not available") - return self._api - - def make_function(self, msg, i, multipart, async): - if (async): - def f(**kwargs): - return self._call_vpp_async(i, msg, **kwargs) - else: - def f(**kwargs): - return self._call_vpp(i, msg, multipart, **kwargs) - - f.__name__ = str(msg.name) - f.__doc__ = ", ".join(["%s %s" % - (msg.fieldtypes[j], k) for j, k in enumerate(msg.fields)]) - return f - - def _register_functions(self, async=False): - self.id_names = [None] * (self.vpp_dictionary_maxid + 1) - self.id_msgdef = [None] * (self.vpp_dictionary_maxid + 1) - self._api = Empty() - for name, msg in vpp_iterator(self.messages): - n = name + '_' + msg.crc[2:] - i = vpp_api.vac_get_msg_index(n.encode()) - if i > 0: - self.id_msgdef[i] = msg - self.id_names[i] = name - # TODO: Fix multipart (use services) - multipart = True if name.find('_dump') > 0 else False - f = self.make_function(msg, i, multipart, async) - setattr(self._api, name, FuncWrapper(f)) - else: - self.logger.debug( - 'No such message type or failed CRC checksum: %s', n) - - def _write_new_cffi(self, buf): - """Send a binary-packed message to VPP.""" - if not self.connected: - raise IOError(1, 'Not connected') - return vpp_api.vac_write(ffi.from_buffer(buf), len(buf)) - - def _write_legacy_cffi(self, buf): - """Send a binary-packed message to VPP.""" - if not self.connected: - raise IOError(1, 'Not connected') - return vpp_api.vac_write(bytes(buf), len(buf)) - - def _read(self): - if not self.connected: - raise IOError(1, 'Not connected') - mem = ffi.new("char **") - size = ffi.new("int *") - rv = vpp_api.vac_read(mem, size, self.read_timeout) - if rv: - raise IOError(rv, 'vac_read failed') - msg = bytes(ffi.buffer(mem[0], size[0])) - vpp_api.vac_free(mem[0]) - return msg - - def connect_internal(self, name, msg_handler, chroot_prefix, rx_qlen, - async): - pfx = chroot_prefix.encode() if chroot_prefix else ffi.NULL - rv = vpp_api.vac_connect(name.encode(), pfx, msg_handler, rx_qlen) - if rv != 0: - raise IOError(2, 'Connect failed') - self.connected = True - self.vpp_dictionary_maxid = vpp_api.vac_msg_table_max_index() - self._register_functions(async=async) - - # Initialise control ping - crc = self.messages['control_ping'].crc - self.control_ping_index = vpp_api.vac_get_msg_index( - ('control_ping' + '_' + crc[2:]).encode()) - self.control_ping_msgdef = self.messages['control_ping'] - if self.async_thread: - self.event_thread = threading.Thread( - target=self.thread_msg_handler) - self.event_thread.daemon = True - self.event_thread.start() - return rv - - def connect(self, name, chroot_prefix=None, async=False, rx_qlen=32): - """Attach to VPP. - - name - the name of the client. - chroot_prefix - if VPP is chroot'ed, the prefix of the jail - async - if true, messages are sent without waiting for a reply - rx_qlen - the length of the VPP message receive queue between - client and server. - """ - msg_handler = vac_callback_sync if not async else vac_callback_async - return self.connect_internal(name, msg_handler, chroot_prefix, rx_qlen, - async) - - def connect_sync(self, name, chroot_prefix=None, rx_qlen=32): - """Attach to VPP in synchronous mode. Application must poll for events. - - name - the name of the client. - chroot_prefix - if VPP is chroot'ed, the prefix of the jail - rx_qlen - the length of the VPP message receive queue between - client and server. - """ - - return self.connect_internal(name, ffi.NULL, chroot_prefix, rx_qlen, - async=False) - - def disconnect(self): - """Detach from VPP.""" - rv = vpp_api.vac_disconnect() - self.connected = False - self.message_queue.put("terminate event thread") - return rv - - def msg_handler_sync(self, msg): - """Process an incoming message from VPP in sync mode. - - The message may be a reply or it may be an async notification. - """ - r = self.decode_incoming_msg(msg) - if r is None: - return - - # If we have a context, then use the context to find any - # request waiting for a reply - context = 0 - if hasattr(r, 'context') and r.context > 0: - context = r.context - - if context == 0: - # No context -> async notification that we feed to the callback - self.message_queue.put_nowait(r) - else: - raise IOError(2, 'RPC reply message received in event handler') - - def decode_incoming_msg(self, msg): - if not msg: - self.logger.warning('vpp_api.read failed') - return - - i, ci = self.header.unpack(msg, 0) - if self.id_names[i] == 'rx_thread_exit': - return - - # - # Decode message and returns a tuple. - # - msgobj = self.id_msgdef[i] - if not msgobj: - raise IOError(2, 'Reply message undefined') - - r = msgobj.unpack(msg) - - return r - - def msg_handler_async(self, msg): - """Process a message from VPP in async mode. - - In async mode, all messages are returned to the callback. - """ - r = self.decode_incoming_msg(msg) - if r is None: - return - - msgname = type(r).__name__ - - if self.event_callback: - self.event_callback(msgname, r) - - def _control_ping(self, context): - """Send a ping command.""" - self._call_vpp_async(self.control_ping_index, - self.control_ping_msgdef, - context=context) - - def validate_args(self, msg, kwargs): - d = set(kwargs.keys()) - set(msg.field_by_name.keys()) - if d: - raise ValueError('Invalid argument {} to {}'.format(list(d), msg.name)) - - def _call_vpp(self, i, msg, multipart, **kwargs): - """Given a message, send the message and await a reply. - - msgdef - the message packing definition - i - the message type index - multipart - True if the message returns multiple - messages in return. - context - context number - chosen at random if not - supplied. - The remainder of the kwargs are the arguments to the API call. - - The return value is the message or message array containing - the response. It will raise an IOError exception if there was - no response within the timeout window. - """ - - if 'context' not in kwargs: - context = self.get_context() - kwargs['context'] = context - else: - context = kwargs['context'] - kwargs['_vl_msg_id'] = i - - self.validate_args(msg, kwargs) - b = msg.pack(kwargs) - vpp_api.vac_rx_suspend() - self._write(b) - - if multipart: - # Send a ping after the request - we use its response - # to detect that we have seen all results. - self._control_ping(context) - - # Block until we get a reply. - rl = [] - while (True): - msg = self._read() - if not msg: - raise IOError(2, 'VPP API client: read failed') - r = self.decode_incoming_msg(msg) - msgname = type(r).__name__ - if context not in r or r.context == 0 or context != r.context: - self.message_queue.put_nowait(r) - continue - - if not multipart: - rl = r - break - if msgname == 'control_ping_reply': - break - - rl.append(r) - - vpp_api.vac_rx_resume() - - return rl - - def _call_vpp_async(self, i, msg, **kwargs): - """Given a message, send the message and await a reply. - - msgdef - the message packing definition - i - the message type index - context - context number - chosen at random if not - supplied. - The remainder of the kwargs are the arguments to the API call. - """ - if 'context' not in kwargs: - context = self.get_context() - kwargs['context'] = context - else: - context = kwargs['context'] - kwargs['client_index'] = 0 - kwargs['_vl_msg_id'] = i - b = msg.pack(kwargs) - - self._write(b) - - def register_event_callback(self, callback): - """Register a callback for async messages. - - This will be called for async notifications in sync mode, - and all messages in async mode. In sync mode, replies to - requests will not come here. - - callback is a fn(msg_type_name, msg_type) that will be - called when a message comes in. While this function is - executing, note that (a) you are in a background thread and - may wish to use threading.Lock to protect your datastructures, - and (b) message processing from VPP will stop (so if you take - a long while about it you may provoke reply timeouts or cause - VPP to fill the RX buffer). Passing None will disable the - callback. - """ - self.event_callback = callback - - def thread_msg_handler(self): - """Python thread calling the user registerd message handler. - - This is to emulate the old style event callback scheme. Modern - clients should provide their own thread to poll the event - queue. - """ - while True: - r = self.message_queue.get() - if r == "terminate event thread": - break - msgname = type(r).__name__ - if self.event_callback: - self.event_callback(msgname, r) - - -# vim: tabstop=8 expandtab shiftwidth=4 softtabstop=4 diff --git a/src/vpp-api/python/vpp_papi/vpp_serializer.py b/src/vpp-api/python/vpp_papi/vpp_serializer.py deleted file mode 100644 index 146a8f6919a..00000000000 --- a/src/vpp-api/python/vpp_papi/vpp_serializer.py +++ /dev/null @@ -1,332 +0,0 @@ -# -# Copyright (c) 2018 Cisco and/or its affiliates. -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at: -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# - -import struct -import collections -from enum import IntEnum -import logging - -# -# Set log-level in application by doing e.g.: -# logger = logging.getLogger('vpp_serializer') -# logger.setLevel(logging.DEBUG) -# -logger = logging.getLogger(__name__) -FORMAT = "[%(filename)s:%(lineno)s - %(funcName)s() ] %(message)s" -logging.basicConfig(format=FORMAT) - - -class BaseTypes(): - def __init__(self, type, elements=0): - base_types = {'u8': '>B', - 'u16': '>H', - 'u32': '>I', - 'i32': '>i', - 'u64': '>Q', - 'f64': '>d', - 'header': '>HI'} - - if elements > 0 and type == 'u8': - self.packer = struct.Struct('>%ss' % elements) - else: - self.packer = struct.Struct(base_types[type]) - self.size = self.packer.size - logger.debug('Adding {} with format: {}' - .format(type, base_types[type])) - - def pack(self, data, kwargs=None): - logger.debug("Data: {} Format: {}".format(data, self.packer.format)) - return self.packer.pack(data) - - def unpack(self, data, offset, result=None): - logger.debug("@ {} Format: {}".format(offset, self.packer.format)) - return self.packer.unpack_from(data, offset)[0] - - -types = {} -types['u8'] = BaseTypes('u8') -types['u16'] = BaseTypes('u16') -types['u32'] = BaseTypes('u32') -types['i32'] = BaseTypes('i32') -types['u64'] = BaseTypes('u64') -types['f64'] = BaseTypes('f64') - - -class FixedList_u8(): - def __init__(self, name, field_type, num): - self.name = name - self.num = num - self.packer = BaseTypes(field_type, num) - self.size = self.packer.size - - def pack(self, list, kwargs): - logger.debug("Data: {}".format(list)) - - if len(list) > self.num: - raise ValueError('Fixed list length error for "{}", got: {}' - ' expected: {}' - .format(self.name, len(list), self.num)) - return self.packer.pack(list) - - def unpack(self, data, offset=0, result=None): - if len(data[offset:]) < self.num: - raise ValueError('Invalid array length for "{}" got {}' - ' expected {}' - .format(self.name, len(data), self.num)) - return self.packer.unpack(data, offset) - - -class FixedList(): - def __init__(self, name, field_type, num): - self.num = num - self.packer = types[field_type] - self.size = self.packer.size * num - - def pack(self, list, kwargs): - logger.debug("Data: {}".format(list)) - - if len(list) != self.num: - raise ValueError('Fixed list length error, got: {} expected: {}' - .format(len(list), self.num)) - b = bytes() - for e in list: - b += self.packer.pack(e) - return b - - def unpack(self, data, offset=0, result=None): - # Return a list of arguments - result = [] - for e in range(self.num): - x = self.packer.unpack(data, offset) - result.append(x) - offset += self.packer.size - return result - - -class VLAList(): - def __init__(self, name, field_type, len_field_name, index): - self.index = index - self.packer = types[field_type] - self.size = self.packer.size - self.length_field = len_field_name - - def pack(self, list, kwargs=None): - logger.debug("Data: {}".format(list)) - if len(list) != kwargs[self.length_field]: - raise ValueError('Variable length error, got: {} expected: {}' - .format(len(list), kwargs[self.length_field])) - b = bytes() - - # u8 array - if self.packer.size == 1: - p = BaseTypes('u8', len(list)) - return p.pack(list) - - for e in list: - b += self.packer.pack(e) - return b - - def unpack(self, data, offset=0, result=None): - logger.debug("Data: {} @ {} Result: {}" - .format(list, offset, result[self.index])) - # Return a list of arguments - - # u8 array - if self.packer.size == 1: - if result[self.index] == 0: - return b'' - p = BaseTypes('u8', result[self.index]) - r = p.unpack(data, offset) - return r - - r = [] - for e in range(result[self.index]): - x = self.packer.unpack(data, offset) - r.append(x) - offset += self.packer.size - return r - - -class VLAList_legacy(): - def __init__(self, name, field_type): - self.packer = types[field_type] - self.size = self.packer.size - - def pack(self, list, kwargs=None): - logger.debug("Data: {}".format(list)) - b = bytes() - for e in list: - b += self.packer.pack(e) - return b - - def unpack(self, data, offset=0, result=None): - # Return a list of arguments - if (len(data) - offset) % self.packer.size: - raise ValueError('Legacy Variable Length Array length mismatch.') - elements = int((len(data) - offset) / self.packer.size) - r = [] - logger.debug("Legacy VLA: {} elements of size {}" - .format(elements, self.packer.size)) - for e in range(elements): - x = self.packer.unpack(data, offset) - r.append(x) - offset += self.packer.size - return r - - -class VPPEnumType(): - def __init__(self, name, msgdef): - self.size = types['u32'].size - e_hash = {} - for f in msgdef: - if type(f) is dict and 'enumtype' in f: - if f['enumtype'] != 'u32': - raise NotImplementedError - continue - ename, evalue = f - e_hash[ename] = evalue - self.enum = IntEnum(name, e_hash) - types[name] = self - logger.debug('Adding enum {}'.format(name)) - - def __getattr__(self, name): - return self.enum[name] - - def pack(self, data, kwargs=None): - logger.debug("Data: {}".format(data)) - return types['u32'].pack(data, kwargs) - - def unpack(self, data, offset=0, result=None): - x = types['u32'].unpack(data, offset) - return self.enum(x) - - -class VPPUnionType(): - def __init__(self, name, msgdef): - self.name = name - self.size = 0 - self.maxindex = 0 - fields = [] - self.packers = collections.OrderedDict() - for i, f in enumerate(msgdef): - if type(f) is dict and 'crc' in f: - self.crc = f['crc'] - continue - f_type, f_name = f - if f_type not in types: - logger.debug('Unknown union type {}'.format(f_type)) - raise ValueError('Unknown message type {}'.format(f_type)) - fields.append(f_name) - size = types[f_type].size - self.packers[f_name] = types[f_type] - if size > self.size: - self.size = size - self.maxindex = i - - types[name] = self - self.tuple = collections.namedtuple(name, fields, rename=True) - logger.debug('Adding union {}'.format(name)) - - def pack(self, data, kwargs=None): - logger.debug("Data: {}".format(data)) - for k, v in data.items(): - logger.debug("Key: {} Value: {}".format(k, v)) - b = self.packers[k].pack(v, kwargs) - offset = self.size - self.packers[k].size - break - r = bytearray(self.size) - r[offset:] = b - return r - - def unpack(self, data, offset=0, result=None): - r = [] - for k, p in self.packers.items(): - union_offset = self.size - p.size - r.append(p.unpack(data, offset + union_offset)) - return self.tuple._make(r) - - -class VPPType(): - # Set everything up to be able to pack / unpack - def __init__(self, name, msgdef): - self.name = name - self.msgdef = msgdef - self.packers = [] - self.fields = [] - self.fieldtypes = [] - self.field_by_name = {} - size = 0 - for i, f in enumerate(msgdef): - if type(f) is dict and 'crc' in f: - self.crc = f['crc'] - continue - f_type, f_name = f[:2] - self.fields.append(f_name) - self.field_by_name[f_name] = None - self.fieldtypes.append(f_type) - if f_type not in types: - logger.debug('Unknown type {}'.format(f_type)) - raise ValueError('Unknown message type {}'.format(f_type)) - if len(f) == 3: # list - list_elements = f[2] - if list_elements == 0: - p = VLAList_legacy(f_name, f_type) - self.packers.append(p) - elif f_type == 'u8': - p = FixedList_u8(f_name, f_type, list_elements) - self.packers.append(p) - size += p.size - else: - p = FixedList(f_name, f_type, list_elements) - self.packers.append(p) - size += p.size - elif len(f) == 4: # Variable length list - # Find index of length field - length_index = self.fields.index(f[3]) - p = VLAList(f_name, f_type, f[3], length_index) - self.packers.append(p) - else: - self.packers.append(types[f_type]) - size += types[f_type].size - - self.size = size - self.tuple = collections.namedtuple(name, self.fields, rename=True) - types[name] = self - logger.debug('Adding type {}'.format(name)) - - def pack(self, data, kwargs=None): - if not kwargs: - kwargs = data - logger.debug("Data: {}".format(data)) - b = bytes() - for i, a in enumerate(self.fields): - if a not in data: - logger.debug("Argument {} not given, defaulting to 0" - .format(a)) - b += b'\x00' * self.packers[i].size - continue - b += self.packers[i].pack(data[a], kwargs) - return b - - def unpack(self, data, offset=0, result=None): - # Return a list of arguments - result = [] - for p in self.packers: - x = p.unpack(data, offset, result) - if type(x) is tuple and len(x) == 1: - x = x[0] - result.append(x) - offset += p.size - return self.tuple._make(result) |