aboutsummaryrefslogtreecommitdiffstats
path: root/src/vpp-api
diff options
context:
space:
mode:
Diffstat (limited to 'src/vpp-api')
-rw-r--r--src/vpp-api/python/setup.py11
-rwxr-xr-xsrc/vpp-api/python/tests/test_vpp_serializer.py114
-rw-r--r--src/vpp-api/python/vpp_papi.py (renamed from src/vpp-api/python/vpp_papi/vpp_papi.py)393
-rw-r--r--src/vpp-api/python/vpp_papi/__init__.py1
-rw-r--r--src/vpp-api/python/vpp_papi/vpp_serializer.py332
5 files changed, 309 insertions, 542 deletions
diff --git a/src/vpp-api/python/setup.py b/src/vpp-api/python/setup.py
index 20b9901d521..abda43de606 100644
--- a/src/vpp-api/python/setup.py
+++ b/src/vpp-api/python/setup.py
@@ -13,20 +13,21 @@
# limitations under the License.
try:
- from setuptools import setup, find_packages
+ from setuptools import setup
except ImportError:
- from distutils.core import setup, find_packages
+ from distutils.core import setup
setup (name = 'vpp_papi',
- version = '1.5',
+ version = '1.4',
description = 'VPP Python binding',
author = 'Ole Troan',
author_email = 'ot@cisco.com',
url = 'https://wiki.fd.io/view/VPP/Python_API',
+ python_requires='>=2.7, >=3.3',
license = 'Apache-2.0',
test_suite = 'tests',
- install_requires=['cffi >= 1.6', 'enum34'],
- packages=find_packages(),
+ install_requires=['cffi >= 1.6'],
+ py_modules=['vpp_papi'],
long_description = '''VPP Python language binding.''',
zip_safe = True,
)
diff --git a/src/vpp-api/python/tests/test_vpp_serializer.py b/src/vpp-api/python/tests/test_vpp_serializer.py
deleted file mode 100755
index 6b867f9e6fe..00000000000
--- a/src/vpp-api/python/tests/test_vpp_serializer.py
+++ /dev/null
@@ -1,114 +0,0 @@
-#!/usr/bin/env python3
-
-import unittest
-from vpp_serializer import VPPType, VPPEnumType, VPPUnionType
-from socket import inet_pton, AF_INET, AF_INET6
-import logging
-
-logging.basicConfig(format='%(levelname)s:%(message)s', level=logging.DEBUG)
-es_logger = logging.getLogger('vpp_serializer')
-es_logger.setLevel(logging.DEBUG)
-
-class VPPMessage(VPPType):
- pass
-
-
-class TestAddType(unittest.TestCase):
-
- def test_union(self):
- un = VPPUnionType('test_union',
- [['u8', 'is_bool'],
- ['u32', 'is_int']])
-
- b = un.pack({'is_int': 0x1234})
- self.assertEqual(len(b), 4)
- nt = un.unpack(b)
- self.assertEqual(nt.is_bool, 52)
- self.assertEqual(nt.is_int, 0x1234)
-
- def test_address(self):
- af = VPPEnumType('vl_api_address_family_t', [["ADDRESS_IP4", 0],
- ["ADDRESS_IP6", 1],
- {"enumtype": "u32"}])
- ip4 = VPPType('vl_api_ip4_address_t', [['u8', 'address', 4]])
- ip6 = VPPType('vl_api_ip6_address_t', [['u8', 'address', 16]])
- VPPUnionType('vl_api_address_union_t',
- [["vl_api_ip4_address_t", "ip4"],
- ["vl_api_ip6_address_t", "ip6"]])
-
- address = VPPType('address', [['vl_api_address_family_t', 'af'],
- ['vl_api_address_union_t', 'un']])
-
- b = ip4.pack({'address': inet_pton(AF_INET, '1.1.1.1')})
- self.assertEqual(len(b), 4)
- nt = ip4.unpack(b)
- self.assertEqual(nt.address, inet_pton(AF_INET, '1.1.1.1'))
-
- b = ip6.pack({'address': inet_pton(AF_INET6, '1::1')})
- self.assertEqual(len(b), 16)
-
- b = address.pack({'af': af.ADDRESS_IP4,
- 'un':
- {'ip4':
- {'address': inet_pton(AF_INET, '2.2.2.2')}}})
- self.assertEqual(len(b), 20)
-
- nt = address.unpack(b)
- print('NT union', nt)
- self.assertEqual(nt.af, af.ADDRESS_IP4)
- self.assertEqual(nt.un.ip4.address,
- inet_pton(AF_INET, '2.2.2.2'))
- self.assertEqual(nt.un.ip6.address,
- inet_pton(AF_INET6, '::0202:0202'))
-
- def test_arrays(self):
- # Test cases
- # 1. Fixed list
- # 2. Fixed list of variable length sub type
- # 3. Variable length type
- #
- ip4 = VPPType('ip4_address', [['u8', 'address', 4]])
- listip4 = VPPType('list_ip4_t', [['ip4_address', 'addresses', 4]])
- valistip4 = VPPType('list_ip4_t',
- [['u8', 'count'],
- ['ip4_address', 'addresses', 0, 'count']])
-
- valistip4_legacy = VPPType('list_ip4_t',
- [['u8', 'foo'],
- ['ip4_address', 'addresses', 0]])
-
- addresses = []
- for i in range(4):
- addresses.append({'address': inet_pton(AF_INET, '2.2.2.2')})
- b = listip4.pack({'addresses': addresses})
- self.assertEqual(len(b), 16)
- nt = listip4.unpack(b)
-
- self.assertEqual(nt.addresses[0].address,
- inet_pton(AF_INET, '2.2.2.2'))
-
- b = valistip4.pack({'count': len(addresses), 'addresses': addresses})
- self.assertEqual(len(b), 17)
-
- nt = valistip4.unpack(b)
- print('NT', nt)
-
- b = valistip4_legacy.pack({'foo': 1, 'addresses': addresses})
- self.assertEqual(len(b), 17)
- nt = valistip4_legacy.unpack(b)
- print('NT', nt)
-
-
- def test_message(self):
- foo = VPPMessage('foo', [['u16', '_vl_msg_id'],
- ['u8', 'client_index'],
- ['u8', 'something'],
- {"crc": "0x559b9f3c"}])
- b = foo.pack({'_vl_msg_id': 1, 'client_index': 5,
- 'something': 200})
- self.assertEqual(len(b), 4)
- nt = foo.unpack(b)
- print('NT', nt)
-
-if __name__ == '__main__':
- unittest.main()
diff --git a/src/vpp-api/python/vpp_papi/vpp_papi.py b/src/vpp-api/python/vpp_papi.py
index 5ff8064d425..ece0e4ee52d 100644
--- a/src/vpp-api/python/vpp_papi/vpp_papi.py
+++ b/src/vpp-api/python/vpp_papi.py
@@ -27,7 +27,6 @@ import weakref
import atexit
from cffi import FFI
import cffi
-from vpp_serializer import VPPType, VPPEnumType, VPPUnionType, BaseTypes
if sys.version[0] == '2':
import Queue as queue
@@ -104,9 +103,6 @@ class FuncWrapper(object):
return self._func(**kwargs)
-class VPPMessage(VPPType):
- pass
-
class VPP():
"""VPP interface.
@@ -119,56 +115,8 @@ class VPP():
provides a means to register a callback function to receive
these messages in a background thread.
"""
-
- def process_json_file(self, apidef_file):
- api = json.load(apidef_file)
- types = {}
- for t in api['enums']:
- t[0] = 'vl_api_' + t[0] + '_t'
- types[t[0]] = {'type': 'enum', 'data': t}
- for t in api['unions']:
- t[0] = 'vl_api_' + t[0] + '_t'
- types[t[0]] = {'type': 'union', 'data': t}
- for t in api['types']:
- t[0] = 'vl_api_' + t[0] + '_t'
- types[t[0]] = {'type': 'type', 'data': t}
-
- i = 0
- while True:
- unresolved = {}
- for k, v in types.items():
- t = v['data']
- if v['type'] == 'enum':
- try:
- VPPEnumType(t[0], t[1:])
- except ValueError:
- unresolved[k] = v
- elif v['type'] == 'union':
- try:
- VPPUnionType(t[0], t[1:])
- except ValueError:
- unresolved[k] = v
- elif v['type'] == 'type':
- try:
- VPPType(t[0], t[1:])
- except ValueError:
- unresolved[k] = v
- if len(unresolved) == 0:
- break
- if i > 3:
- raise ValueError('Unresolved type definitions {}'
- .format(unresolved))
- types = unresolved
- i += 1
-
- for m in api['messages']:
- try:
- self.messages[m[0]] = VPPMessage(m[0], m[1:])
- except NotImplementedError:
- self.logger.error('Not implemented error for {}'.format(m[0]))
-
def __init__(self, apifiles=None, testmode=False, async_thread=True,
- logger=logging.getLogger('vpp_papi'), loglevel='debug',
+ logger=None, loglevel=None,
read_timeout=0):
"""Create a VPP API object.
@@ -189,14 +137,14 @@ class VPP():
logger = logging.getLogger(__name__)
if loglevel is not None:
logger.setLevel(loglevel)
+
self.logger = logger
self.messages = {}
self.id_names = []
self.id_msgdef = []
self.connected = False
- self.header = VPPType('header', [['u16', 'msgid'],
- ['u32', 'client_index']])
+ self.header = struct.Struct('>HI')
self.apifiles = []
self.event_callback = None
self.message_queue = queue.Queue()
@@ -217,8 +165,12 @@ class VPP():
for file in apifiles:
with open(file) as apidef_file:
- self.process_json_file(apidef_file)
+ api = json.load(apidef_file)
+ for t in api['types']:
+ self.add_type(t[0], t[1:])
+ for m in api['messages']:
+ self.add_message(m[0], m[1:])
self.apifiles = apifiles
# Basic sanity check
@@ -229,8 +181,7 @@ class VPP():
atexit.register(vpp_atexit, weakref.ref(self))
# Register error handler
- if not testmode:
- vpp_api.vac_set_error_handler(vac_error_handler)
+ vpp_api.vac_set_error_handler(vac_error_handler)
# Support legacy CFFI
# from_buffer supported from 1.8.0
@@ -383,38 +334,305 @@ class VPP():
print('Connected') if self.connected else print('Not Connected')
print('Read API definitions from', ', '.join(self.apifiles))
- @property
- def api(self):
- if not hasattr(self, "_api"):
- raise Exception("Not connected, api definitions not available")
- return self._api
+ def __struct(self, t, n=None, e=-1, vl=None):
+ """Create a packing structure for a message."""
+ base_types = {'u8': 'B',
+ 'u16': 'H',
+ 'u32': 'I',
+ 'i32': 'i',
+ 'u64': 'Q',
+ 'f64': 'd', }
+
+ if t in base_types:
+ if not vl:
+ if e > 0 and t == 'u8':
+ # Fixed byte array
+ s = struct.Struct('>' + str(e) + 's')
+ return s.size, s
+ if e > 0:
+ # Fixed array of base type
+ s = struct.Struct('>' + base_types[t])
+ return s.size, [e, s]
+ elif e == 0:
+ # Old style variable array
+ s = struct.Struct('>' + base_types[t])
+ return s.size, [-1, s]
+ else:
+ # Variable length array
+ if t == 'u8':
+ s = struct.Struct('>s')
+ return s.size, [vl, s]
+ else:
+ s = struct.Struct('>' + base_types[t])
+ return s.size, [vl, s]
+
+ s = struct.Struct('>' + base_types[t])
+ return s.size, s
+
+ if t in self.messages:
+ size = self.messages[t]['sizes'][0]
+
+ # Return a list in case of array
+ if e > 0 and not vl:
+ return size, [e, lambda self, encode, buf, offset, args: (
+ self.__struct_type(encode, self.messages[t], buf, offset,
+ args))]
+ if vl:
+ return size, [vl, lambda self, encode, buf, offset, args: (
+ self.__struct_type(encode, self.messages[t], buf, offset,
+ args))]
+ elif e == 0:
+ # Old style VLA
+ raise NotImplementedError(1,
+ 'No support for compound types ' + t)
+ return size, lambda self, encode, buf, offset, args: (
+ self.__struct_type(encode, self.messages[t], buf, offset, args)
+ )
+
+ raise ValueError(1, 'Invalid message type: ' + t)
+
+ def __struct_type(self, encode, msgdef, buf, offset, kwargs):
+ """Get a message packer or unpacker."""
+ if encode:
+ return self.__struct_type_encode(msgdef, buf, offset, kwargs)
+ else:
+ return self.__struct_type_decode(msgdef, buf, offset)
+
+ def __struct_type_encode(self, msgdef, buf, offset, kwargs):
+ off = offset
+ size = 0
+
+ for k in kwargs:
+ if k not in msgdef['args']:
+ raise ValueError(1, 'Non existing argument [' + k + ']' +
+ ' used in call to: ' +
+ self.id_names[kwargs['_vl_msg_id']] + '()')
+
+ for k, v in vpp_iterator(msgdef['args']):
+ off += size
+ if k in kwargs:
+ if type(v) is list:
+ if callable(v[1]):
+ e = kwargs[v[0]] if v[0] in kwargs else v[0]
+ if e != len(kwargs[k]):
+ raise (ValueError(1,
+ 'Input list length mismatch: '
+ '%s (%s != %s)' %
+ (k, e, len(kwargs[k]))))
+ size = 0
+ for i in range(e):
+ size += v[1](self, True, buf, off + size,
+ kwargs[k][i])
+ else:
+ if v[0] in kwargs:
+ kwargslen = kwargs[v[0]]
+ if kwargslen != len(kwargs[k]):
+ raise ValueError(1,
+ 'Input list length mismatch:'
+ ' %s (%s != %s)' %
+ (k, kwargslen,
+ len(kwargs[k])))
+ else:
+ kwargslen = len(kwargs[k])
+ if v[1].size == 1:
+ buf[off:off + kwargslen] = bytearray(kwargs[k])
+ size = kwargslen
+ else:
+ size = 0
+ for i in kwargs[k]:
+ v[1].pack_into(buf, off + size, i)
+ size += v[1].size
+ else:
+ if callable(v):
+ size = v(self, True, buf, off, kwargs[k])
+ else:
+ if type(kwargs[k]) is str and v.size < len(kwargs[k]):
+ raise ValueError(1,
+ 'Input list length mismatch: '
+ '%s (%s < %s)' %
+ (k, v.size, len(kwargs[k])))
+ v.pack_into(buf, off, kwargs[k])
+ size = v.size
+ else:
+ size = v.size if not type(v) is list else 0
+
+ return off + size - offset
+
+ def __getitem__(self, name):
+ if name in self.messages:
+ return self.messages[name]
+ return None
- def make_function(self, msg, i, multipart, async):
+ def get_size(self, sizes, kwargs):
+ total_size = sizes[0]
+ for e in sizes[1]:
+ if e in kwargs and type(kwargs[e]) is list:
+ total_size += len(kwargs[e]) * sizes[1][e]
+ return total_size
+
+ def encode(self, msgdef, kwargs):
+ # Make suitably large buffer
+ size = self.get_size(msgdef['sizes'], kwargs)
+ buf = bytearray(size)
+ offset = 0
+ size = self.__struct_type(True, msgdef, buf, offset, kwargs)
+ return buf[:offset + size]
+
+ def decode(self, msgdef, buf):
+ return self.__struct_type(False, msgdef, buf, 0, None)[1]
+
+ def __struct_type_decode(self, msgdef, buf, offset):
+ res = []
+ off = offset
+ size = 0
+ for k, v in vpp_iterator(msgdef['args']):
+ off += size
+ if type(v) is list:
+ lst = []
+ if callable(v[1]): # compound type
+ size = 0
+ if v[0] in msgdef['args']: # vla
+ e = res[v[2]]
+ else: # fixed array
+ e = v[0]
+ res.append(lst)
+ for i in range(e):
+ (s, l) = v[1](self, False, buf, off + size, None)
+ lst.append(l)
+ size += s
+ continue
+ if v[1].size == 1:
+ if type(v[0]) is int:
+ size = len(buf) - off
+ else:
+ size = res[v[2]]
+ res.append(buf[off:off + size])
+ else:
+ e = v[0] if type(v[0]) is int else res[v[2]]
+ if e == -1:
+ e = (len(buf) - off) / v[1].size
+ lst = []
+ res.append(lst)
+ size = 0
+ for i in range(e):
+ lst.append(v[1].unpack_from(buf, off + size)[0])
+ size += v[1].size
+ else:
+ if callable(v):
+ size = 0
+ (s, l) = v(self, False, buf, off, None)
+ res.append(l)
+ size += s
+ else:
+ res.append(v.unpack_from(buf, off)[0])
+ size = v.size
+
+ return off + size - offset, msgdef['return_tuple']._make(res)
+
+ def ret_tup(self, name):
+ if name in self.messages and 'return_tuple' in self.messages[name]:
+ return self.messages[name]['return_tuple']
+ return None
+
+ def duplicate_check_ok(self, name, msgdef):
+ crc = None
+ for c in msgdef:
+ if type(c) is dict and 'crc' in c:
+ crc = c['crc']
+ break
+ if crc:
+ # We can get duplicates because of imports
+ if crc == self.messages[name]['crc']:
+ return True
+ return False
+
+ def add_message(self, name, msgdef, typeonly=False):
+ if name in self.messages:
+ if typeonly:
+ if self.duplicate_check_ok(name, msgdef):
+ return
+ raise ValueError('Duplicate message name: ' + name)
+
+ args = collections.OrderedDict()
+ argtypes = collections.OrderedDict()
+ fields = []
+ msg = {}
+ total_size = 0
+ sizes = {}
+ for i, f in enumerate(msgdef):
+ if type(f) is dict and 'crc' in f:
+ msg['crc'] = f['crc']
+ continue
+ field_type = f[0]
+ field_name = f[1]
+ if len(f) == 3 and f[2] == 0 and i != len(msgdef) - 2:
+ raise ValueError('Variable Length Array must be last: ' + name)
+ size, s = self.__struct(*f)
+ args[field_name] = s
+ if type(s) == list and type(s[0]) == int and \
+ type(s[1]) == struct.Struct:
+ if s[0] < 0:
+ sizes[field_name] = size
+ else:
+ sizes[field_name] = size
+ total_size += s[0] * size
+ else:
+ sizes[field_name] = size
+ total_size += size
+
+ argtypes[field_name] = field_type
+ if len(f) == 4: # Find offset to # elements field
+ idx = list(args.keys()).index(f[3]) - i
+ args[field_name].append(idx)
+ fields.append(field_name)
+ msg['return_tuple'] = collections.namedtuple(name, fields,
+ rename=True)
+ self.messages[name] = msg
+ self.messages[name]['args'] = args
+ self.messages[name]['argtypes'] = argtypes
+ self.messages[name]['typeonly'] = typeonly
+ self.messages[name]['sizes'] = [total_size, sizes]
+ return self.messages[name]
+
+ def add_type(self, name, typedef):
+ return self.add_message('vl_api_' + name + '_t', typedef,
+ typeonly=True)
+
+ def make_function(self, name, i, msgdef, multipart, async):
if (async):
def f(**kwargs):
- return self._call_vpp_async(i, msg, **kwargs)
+ return self._call_vpp_async(i, msgdef, **kwargs)
else:
def f(**kwargs):
- return self._call_vpp(i, msg, multipart, **kwargs)
-
- f.__name__ = str(msg.name)
+ return self._call_vpp(i, msgdef, multipart, **kwargs)
+ args = self.messages[name]['args']
+ argtypes = self.messages[name]['argtypes']
+ f.__name__ = str(name)
f.__doc__ = ", ".join(["%s %s" %
- (msg.fieldtypes[j], k) for j, k in enumerate(msg.fields)])
+ (argtypes[k], k) for k in args.keys()])
return f
+ @property
+ def api(self):
+ if not hasattr(self, "_api"):
+ raise Exception("Not connected, api definitions not available")
+ return self._api
+
def _register_functions(self, async=False):
self.id_names = [None] * (self.vpp_dictionary_maxid + 1)
self.id_msgdef = [None] * (self.vpp_dictionary_maxid + 1)
self._api = Empty()
- for name, msg in vpp_iterator(self.messages):
- n = name + '_' + msg.crc[2:]
+ for name, msgdef in vpp_iterator(self.messages):
+ if self.messages[name]['typeonly']:
+ continue
+ crc = self.messages[name]['crc']
+ n = name + '_' + crc[2:]
i = vpp_api.vac_get_msg_index(n.encode())
if i > 0:
- self.id_msgdef[i] = msg
+ self.id_msgdef[i] = msgdef
self.id_names[i] = name
- # TODO: Fix multipart (use services)
multipart = True if name.find('_dump') > 0 else False
- f = self.make_function(msg, i, multipart, async)
+ f = self.make_function(name, i, msgdef, multipart, async)
setattr(self._api, name, FuncWrapper(f))
else:
self.logger.debug(
@@ -451,11 +669,12 @@ class VPP():
if rv != 0:
raise IOError(2, 'Connect failed')
self.connected = True
+
self.vpp_dictionary_maxid = vpp_api.vac_msg_table_max_index()
self._register_functions(async=async)
# Initialise control ping
- crc = self.messages['control_ping'].crc
+ crc = self.messages['control_ping']['crc']
self.control_ping_index = vpp_api.vac_get_msg_index(
('control_ping' + '_' + crc[2:]).encode())
self.control_ping_msgdef = self.messages['control_ping']
@@ -524,18 +743,18 @@ class VPP():
self.logger.warning('vpp_api.read failed')
return
- i, ci = self.header.unpack(msg, 0)
+ i, ci = self.header.unpack_from(msg, 0)
if self.id_names[i] == 'rx_thread_exit':
return
#
# Decode message and returns a tuple.
#
- msgobj = self.id_msgdef[i]
- if not msgobj:
+ msgdef = self.id_msgdef[i]
+ if not msgdef:
raise IOError(2, 'Reply message undefined')
- r = msgobj.unpack(msg)
+ r = self.decode(msgdef, msg)
return r
@@ -559,12 +778,7 @@ class VPP():
self.control_ping_msgdef,
context=context)
- def validate_args(self, msg, kwargs):
- d = set(kwargs.keys()) - set(msg.field_by_name.keys())
- if d:
- raise ValueError('Invalid argument {} to {}'.format(list(d), msg.name))
-
- def _call_vpp(self, i, msg, multipart, **kwargs):
+ def _call_vpp(self, i, msgdef, multipart, **kwargs):
"""Given a message, send the message and await a reply.
msgdef - the message packing definition
@@ -586,9 +800,8 @@ class VPP():
else:
context = kwargs['context']
kwargs['_vl_msg_id'] = i
+ b = self.encode(msgdef, kwargs)
- self.validate_args(msg, kwargs)
- b = msg.pack(kwargs)
vpp_api.vac_rx_suspend()
self._write(b)
@@ -603,6 +816,7 @@ class VPP():
msg = self._read()
if not msg:
raise IOError(2, 'VPP API client: read failed')
+
r = self.decode_incoming_msg(msg)
msgname = type(r).__name__
if context not in r or r.context == 0 or context != r.context:
@@ -621,7 +835,7 @@ class VPP():
return rl
- def _call_vpp_async(self, i, msg, **kwargs):
+ def _call_vpp_async(self, i, msgdef, **kwargs):
"""Given a message, send the message and await a reply.
msgdef - the message packing definition
@@ -635,9 +849,8 @@ class VPP():
kwargs['context'] = context
else:
context = kwargs['context']
- kwargs['client_index'] = 0
kwargs['_vl_msg_id'] = i
- b = msg.pack(kwargs)
+ b = self.encode(msgdef, kwargs)
self._write(b)
diff --git a/src/vpp-api/python/vpp_papi/__init__.py b/src/vpp-api/python/vpp_papi/__init__.py
deleted file mode 100644
index f9afcf17f29..00000000000
--- a/src/vpp-api/python/vpp_papi/__init__.py
+++ /dev/null
@@ -1 +0,0 @@
-from .vpp_papi import *
diff --git a/src/vpp-api/python/vpp_papi/vpp_serializer.py b/src/vpp-api/python/vpp_papi/vpp_serializer.py
deleted file mode 100644
index 146a8f6919a..00000000000
--- a/src/vpp-api/python/vpp_papi/vpp_serializer.py
+++ /dev/null
@@ -1,332 +0,0 @@
-#
-# Copyright (c) 2018 Cisco and/or its affiliates.
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at:
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#
-
-import struct
-import collections
-from enum import IntEnum
-import logging
-
-#
-# Set log-level in application by doing e.g.:
-# logger = logging.getLogger('vpp_serializer')
-# logger.setLevel(logging.DEBUG)
-#
-logger = logging.getLogger(__name__)
-FORMAT = "[%(filename)s:%(lineno)s - %(funcName)s() ] %(message)s"
-logging.basicConfig(format=FORMAT)
-
-
-class BaseTypes():
- def __init__(self, type, elements=0):
- base_types = {'u8': '>B',
- 'u16': '>H',
- 'u32': '>I',
- 'i32': '>i',
- 'u64': '>Q',
- 'f64': '>d',
- 'header': '>HI'}
-
- if elements > 0 and type == 'u8':
- self.packer = struct.Struct('>%ss' % elements)
- else:
- self.packer = struct.Struct(base_types[type])
- self.size = self.packer.size
- logger.debug('Adding {} with format: {}'
- .format(type, base_types[type]))
-
- def pack(self, data, kwargs=None):
- logger.debug("Data: {} Format: {}".format(data, self.packer.format))
- return self.packer.pack(data)
-
- def unpack(self, data, offset, result=None):
- logger.debug("@ {} Format: {}".format(offset, self.packer.format))
- return self.packer.unpack_from(data, offset)[0]
-
-
-types = {}
-types['u8'] = BaseTypes('u8')
-types['u16'] = BaseTypes('u16')
-types['u32'] = BaseTypes('u32')
-types['i32'] = BaseTypes('i32')
-types['u64'] = BaseTypes('u64')
-types['f64'] = BaseTypes('f64')
-
-
-class FixedList_u8():
- def __init__(self, name, field_type, num):
- self.name = name
- self.num = num
- self.packer = BaseTypes(field_type, num)
- self.size = self.packer.size
-
- def pack(self, list, kwargs):
- logger.debug("Data: {}".format(list))
-
- if len(list) > self.num:
- raise ValueError('Fixed list length error for "{}", got: {}'
- ' expected: {}'
- .format(self.name, len(list), self.num))
- return self.packer.pack(list)
-
- def unpack(self, data, offset=0, result=None):
- if len(data[offset:]) < self.num:
- raise ValueError('Invalid array length for "{}" got {}'
- ' expected {}'
- .format(self.name, len(data), self.num))
- return self.packer.unpack(data, offset)
-
-
-class FixedList():
- def __init__(self, name, field_type, num):
- self.num = num
- self.packer = types[field_type]
- self.size = self.packer.size * num
-
- def pack(self, list, kwargs):
- logger.debug("Data: {}".format(list))
-
- if len(list) != self.num:
- raise ValueError('Fixed list length error, got: {} expected: {}'
- .format(len(list), self.num))
- b = bytes()
- for e in list:
- b += self.packer.pack(e)
- return b
-
- def unpack(self, data, offset=0, result=None):
- # Return a list of arguments
- result = []
- for e in range(self.num):
- x = self.packer.unpack(data, offset)
- result.append(x)
- offset += self.packer.size
- return result
-
-
-class VLAList():
- def __init__(self, name, field_type, len_field_name, index):
- self.index = index
- self.packer = types[field_type]
- self.size = self.packer.size
- self.length_field = len_field_name
-
- def pack(self, list, kwargs=None):
- logger.debug("Data: {}".format(list))
- if len(list) != kwargs[self.length_field]:
- raise ValueError('Variable length error, got: {} expected: {}'
- .format(len(list), kwargs[self.length_field]))
- b = bytes()
-
- # u8 array
- if self.packer.size == 1:
- p = BaseTypes('u8', len(list))
- return p.pack(list)
-
- for e in list:
- b += self.packer.pack(e)
- return b
-
- def unpack(self, data, offset=0, result=None):
- logger.debug("Data: {} @ {} Result: {}"
- .format(list, offset, result[self.index]))
- # Return a list of arguments
-
- # u8 array
- if self.packer.size == 1:
- if result[self.index] == 0:
- return b''
- p = BaseTypes('u8', result[self.index])
- r = p.unpack(data, offset)
- return r
-
- r = []
- for e in range(result[self.index]):
- x = self.packer.unpack(data, offset)
- r.append(x)
- offset += self.packer.size
- return r
-
-
-class VLAList_legacy():
- def __init__(self, name, field_type):
- self.packer = types[field_type]
- self.size = self.packer.size
-
- def pack(self, list, kwargs=None):
- logger.debug("Data: {}".format(list))
- b = bytes()
- for e in list:
- b += self.packer.pack(e)
- return b
-
- def unpack(self, data, offset=0, result=None):
- # Return a list of arguments
- if (len(data) - offset) % self.packer.size:
- raise ValueError('Legacy Variable Length Array length mismatch.')
- elements = int((len(data) - offset) / self.packer.size)
- r = []
- logger.debug("Legacy VLA: {} elements of size {}"
- .format(elements, self.packer.size))
- for e in range(elements):
- x = self.packer.unpack(data, offset)
- r.append(x)
- offset += self.packer.size
- return r
-
-
-class VPPEnumType():
- def __init__(self, name, msgdef):
- self.size = types['u32'].size
- e_hash = {}
- for f in msgdef:
- if type(f) is dict and 'enumtype' in f:
- if f['enumtype'] != 'u32':
- raise NotImplementedError
- continue
- ename, evalue = f
- e_hash[ename] = evalue
- self.enum = IntEnum(name, e_hash)
- types[name] = self
- logger.debug('Adding enum {}'.format(name))
-
- def __getattr__(self, name):
- return self.enum[name]
-
- def pack(self, data, kwargs=None):
- logger.debug("Data: {}".format(data))
- return types['u32'].pack(data, kwargs)
-
- def unpack(self, data, offset=0, result=None):
- x = types['u32'].unpack(data, offset)
- return self.enum(x)
-
-
-class VPPUnionType():
- def __init__(self, name, msgdef):
- self.name = name
- self.size = 0
- self.maxindex = 0
- fields = []
- self.packers = collections.OrderedDict()
- for i, f in enumerate(msgdef):
- if type(f) is dict and 'crc' in f:
- self.crc = f['crc']
- continue
- f_type, f_name = f
- if f_type not in types:
- logger.debug('Unknown union type {}'.format(f_type))
- raise ValueError('Unknown message type {}'.format(f_type))
- fields.append(f_name)
- size = types[f_type].size
- self.packers[f_name] = types[f_type]
- if size > self.size:
- self.size = size
- self.maxindex = i
-
- types[name] = self
- self.tuple = collections.namedtuple(name, fields, rename=True)
- logger.debug('Adding union {}'.format(name))
-
- def pack(self, data, kwargs=None):
- logger.debug("Data: {}".format(data))
- for k, v in data.items():
- logger.debug("Key: {} Value: {}".format(k, v))
- b = self.packers[k].pack(v, kwargs)
- offset = self.size - self.packers[k].size
- break
- r = bytearray(self.size)
- r[offset:] = b
- return r
-
- def unpack(self, data, offset=0, result=None):
- r = []
- for k, p in self.packers.items():
- union_offset = self.size - p.size
- r.append(p.unpack(data, offset + union_offset))
- return self.tuple._make(r)
-
-
-class VPPType():
- # Set everything up to be able to pack / unpack
- def __init__(self, name, msgdef):
- self.name = name
- self.msgdef = msgdef
- self.packers = []
- self.fields = []
- self.fieldtypes = []
- self.field_by_name = {}
- size = 0
- for i, f in enumerate(msgdef):
- if type(f) is dict and 'crc' in f:
- self.crc = f['crc']
- continue
- f_type, f_name = f[:2]
- self.fields.append(f_name)
- self.field_by_name[f_name] = None
- self.fieldtypes.append(f_type)
- if f_type not in types:
- logger.debug('Unknown type {}'.format(f_type))
- raise ValueError('Unknown message type {}'.format(f_type))
- if len(f) == 3: # list
- list_elements = f[2]
- if list_elements == 0:
- p = VLAList_legacy(f_name, f_type)
- self.packers.append(p)
- elif f_type == 'u8':
- p = FixedList_u8(f_name, f_type, list_elements)
- self.packers.append(p)
- size += p.size
- else:
- p = FixedList(f_name, f_type, list_elements)
- self.packers.append(p)
- size += p.size
- elif len(f) == 4: # Variable length list
- # Find index of length field
- length_index = self.fields.index(f[3])
- p = VLAList(f_name, f_type, f[3], length_index)
- self.packers.append(p)
- else:
- self.packers.append(types[f_type])
- size += types[f_type].size
-
- self.size = size
- self.tuple = collections.namedtuple(name, self.fields, rename=True)
- types[name] = self
- logger.debug('Adding type {}'.format(name))
-
- def pack(self, data, kwargs=None):
- if not kwargs:
- kwargs = data
- logger.debug("Data: {}".format(data))
- b = bytes()
- for i, a in enumerate(self.fields):
- if a not in data:
- logger.debug("Argument {} not given, defaulting to 0"
- .format(a))
- b += b'\x00' * self.packers[i].size
- continue
- b += self.packers[i].pack(data[a], kwargs)
- return b
-
- def unpack(self, data, offset=0, result=None):
- # Return a list of arguments
- result = []
- for p in self.packers:
- x = p.unpack(data, offset, result)
- if type(x) is tuple and len(x) == 1:
- x = x[0]
- result.append(x)
- offset += p.size
- return self.tuple._make(result)