mirror of
https://github.com/msgpack/msgpack-python.git
synced 2025-10-22 21:33:17 +00:00
Add pure Python fallback module
Signed-off-by: Bas Westerbaan <bas@westerbaan.name>
This commit is contained in:
parent
5f55e4c6db
commit
6a28b28c63
2 changed files with 482 additions and 2 deletions
476
msgpack/fallback.py
Normal file
476
msgpack/fallback.py
Normal file
|
@ -0,0 +1,476 @@
|
|||
# Fallback pure Python implementation of msgpack
|
||||
|
||||
import sys
|
||||
import array
|
||||
import struct
|
||||
|
||||
try:
|
||||
import cStringIO as StringIO
|
||||
except ImportError:
|
||||
import StringIO
|
||||
|
||||
from msgpack.exceptions import (
|
||||
BufferFull,
|
||||
OutOfData,
|
||||
UnpackValueError,
|
||||
PackValueError,
|
||||
ExtraData)
|
||||
|
||||
EX_SKIP = 0
|
||||
EX_CONSTRUCT = 1
|
||||
EX_READ_ARRAY_HEADER = 2
|
||||
EX_READ_MAP_HEADER = 3
|
||||
|
||||
TYPE_IMMEDIATE = 0
|
||||
TYPE_ARRAY = 1
|
||||
TYPE_MAP = 2
|
||||
TYPE_RAW = 3
|
||||
|
||||
DEFAULT_RECURSE_LIMIT=511
|
||||
|
||||
def pack(o, stream, default=None, encoding='utf-8', unicode_errors='strict'):
|
||||
""" Pack object `o` and write it to `stream` """
|
||||
packer = Packer(default=default, encoding=encoding,
|
||||
unicode_errors=unicode_errors)
|
||||
stream.write(packer.pack(o))
|
||||
|
||||
def packb(o, default=None, encoding='utf-8', unicode_errors='struct',
|
||||
use_single_float=False):
|
||||
""" Pack object `o` and return packed bytes """
|
||||
packer = Packer(default=default,
|
||||
encoding=encoding,
|
||||
unicode_errors=unicode_errors,
|
||||
use_single_float=use_single_float)
|
||||
return packer.pack(o)
|
||||
|
||||
def unpack(stream, object_hook=None, list_hook=None, use_list=True,
|
||||
encoding=None, unicode_errors='strict',
|
||||
object_pairs_hook=None):
|
||||
""" Unpack an object from `stream`.
|
||||
|
||||
Raises `ExtraData` when `stream` has extra bytes. """
|
||||
unpacker = Unpacker(stream, object_hook=object_hook, list_hook=list_hook,
|
||||
use_list=use_list, encoding=encoding, unicode_errors=unicode_errors,
|
||||
object_pairs_hook=object_pairs_hook)
|
||||
ret = unpacker._fb_unpack()
|
||||
unpacker._fb_check_for_extradata()
|
||||
return ret
|
||||
|
||||
def unpackb(packed, object_hook=None, list_hook=None, use_list=True,
|
||||
encoding=None, unicode_errors='strict',
|
||||
object_pairs_hook=None):
|
||||
""" Unpack an object from `packed`.
|
||||
|
||||
Raises `ExtraData` when `packed` contains extra bytes. """
|
||||
unpacker = Unpacker(None, object_hook=object_hook, list_hook=list_hook,
|
||||
use_list=use_list, encoding=encoding, unicode_errors=unicode_errors,
|
||||
object_pairs_hook=object_pairs_hook)
|
||||
unpacker.feed(packed)
|
||||
ret = unpacker._fb_unpack()
|
||||
unpacker._fb_check_for_extradata()
|
||||
return ret
|
||||
|
||||
class Unpacker(object):
|
||||
"""
|
||||
Streaming unpacker.
|
||||
|
||||
`file_like` is a file-like object having a `.read(n)` method.
|
||||
When `Unpacker` is initialized with a `file_like`, `.feed()` is not
|
||||
usable.
|
||||
|
||||
`read_size` is used for `file_like.read(read_size)`.
|
||||
|
||||
If `use_list` is True (default), msgpack lists are deserialized to Python
|
||||
lists. Otherwise they are deserialized to tuples.
|
||||
|
||||
`object_hook` is the same as in simplejson. If it is not None, it should
|
||||
be callable and Unpacker calls it with a dict argument after deserializing
|
||||
a map.
|
||||
|
||||
`object_pairs_hook` is the same as in simplejson. If it is not None, it
|
||||
should be callable and Unpacker calls it with a list of key-value pairs
|
||||
after deserializing a map.
|
||||
|
||||
`encoding` is the encoding used for decoding msgpack bytes. If it is
|
||||
None (default), msgpack bytes are deserialized to Python bytes.
|
||||
|
||||
`unicode_errors` is used for decoding bytes.
|
||||
|
||||
`max_buffer_size` limits the buffer size. 0 means INT_MAX (default).
|
||||
|
||||
Raises `BufferFull` exception when it is unsufficient.
|
||||
|
||||
You should set this parameter when unpacking data from an untrustred source.
|
||||
|
||||
example of streaming deserialization from file-like object::
|
||||
|
||||
unpacker = Unpacker(file_like)
|
||||
for o in unpacker:
|
||||
do_something(o)
|
||||
|
||||
example of streaming deserialization from socket::
|
||||
|
||||
unpacker = Unapcker()
|
||||
while 1:
|
||||
buf = sock.recv(1024*2)
|
||||
if not buf:
|
||||
break
|
||||
unpacker.feed(buf)
|
||||
for o in unpacker:
|
||||
do_something(o)
|
||||
"""
|
||||
|
||||
def __init__(self, file_like=None, read_size=0, use_list=True,
|
||||
object_hook=None, object_pairs_hook=None, list_hook=None,
|
||||
encoding=None, unicode_errors='strict', max_buffer_size=0):
|
||||
if file_like is None:
|
||||
self._fb_feeding = True
|
||||
else:
|
||||
if not callable(file_like.read):
|
||||
raise ValueError("`file_like.read` must be callable")
|
||||
self.file_like = file_like
|
||||
self._fb_feeding = False
|
||||
self._fb_buffers = []
|
||||
self._fb_buf_o = 0
|
||||
self._fb_buf_i = 0
|
||||
self._fb_buf_n = 0
|
||||
self.max_buffer_size = (sys.maxint if max_buffer_size == 0
|
||||
else max_buffer_size)
|
||||
self.read_size = (read_size if read_size != 0
|
||||
else min(self.max_buffer_size, 2048))
|
||||
if read_size > self.max_buffer_size:
|
||||
raise ValueError("read_size must be smaller than max_buffer_size")
|
||||
self.encoding = encoding
|
||||
self.unicode_errors = unicode_errors
|
||||
self.use_list = use_list
|
||||
self.list_hook = list_hook
|
||||
self.object_hook = object_hook
|
||||
self.object_pairs_hook = object_pairs_hook
|
||||
|
||||
if list_hook is not None and not callable(list_hook):
|
||||
raise ValueError('`list_hook` is not callable')
|
||||
if object_hook is not None and not callable(object_hook):
|
||||
raise ValueError('`object_hook` is not callable')
|
||||
if object_pairs_hook is not None and not callable(object_pairs_hook):
|
||||
raise ValueError('`object_pairs_hook` is not callable')
|
||||
if object_hook is not None and object_pairs_hook is not None:
|
||||
raise ValueError("object_pairs_hook and object_hook are mutually "+
|
||||
"exclusive")
|
||||
|
||||
def feed(self, next_bytes):
|
||||
if isinstance(next_bytes, array.array):
|
||||
next_bytes = next_bytes.tostring()
|
||||
assert self._fb_feeding
|
||||
if self._fb_buf_n + len(next_bytes) > self.max_buffer_size:
|
||||
raise BufferFull
|
||||
self._fb_buf_n += len(next_bytes)
|
||||
self._fb_buffers.append(next_bytes)
|
||||
|
||||
def _fb_consume(self):
|
||||
self._fb_buffers = self._fb_buffers[self._fb_buf_i:]
|
||||
if self._fb_buffers:
|
||||
self._fb_buffers[0] = self._fb_buffers[0][self._fb_buf_o:]
|
||||
self._fb_buf_o = 0
|
||||
self._fb_buf_i = 0
|
||||
self._fb_buf_n = sum(map(len, self._fb_buffers))
|
||||
|
||||
def _fb_check_for_extradata(self):
|
||||
if self._fb_buf_i != len(self._fb_buffers):
|
||||
raise ExtraData
|
||||
if self._fb_feeding:
|
||||
return
|
||||
if not self.file_like:
|
||||
return
|
||||
if not self.file_like.read(1):
|
||||
raise ExtraData
|
||||
|
||||
def __iter__(self):
|
||||
return self
|
||||
|
||||
def next(self):
|
||||
try:
|
||||
ret = self._fb_unpack(None, None)
|
||||
self._fb_consume()
|
||||
return ret
|
||||
except OutOfData:
|
||||
raise StopIteration
|
||||
|
||||
def read_bytes(self, n):
|
||||
return self._fb_read(n)
|
||||
|
||||
def _fb_read(self, n, write_bytes=None):
|
||||
ret = ''
|
||||
while len(ret) != n:
|
||||
if self._fb_buf_i == len(self._fb_buffers):
|
||||
if self._fb_feeding:
|
||||
break
|
||||
tmp = self.file_like.read(self.read_size)
|
||||
if not tmp:
|
||||
break
|
||||
self._fb_buffers.append(tmp)
|
||||
continue
|
||||
sliced = n - len(ret)
|
||||
ret += self._fb_buffers[self._fb_buf_i][
|
||||
self._fb_buf_o:self._fb_buf_o + sliced]
|
||||
self._fb_buf_o += sliced
|
||||
if self._fb_buf_o >= len(self._fb_buffers[self._fb_buf_i]):
|
||||
self._fb_buf_o = 0
|
||||
self._fb_buf_i += 1
|
||||
if len(ret) != n:
|
||||
raise OutOfData
|
||||
if write_bytes is not None:
|
||||
write_bytes(ret)
|
||||
return ret
|
||||
|
||||
def _fb_unpack(self, execute=EX_CONSTRUCT, write_bytes=None):
|
||||
typ = TYPE_IMMEDIATE
|
||||
b = ord(self._fb_read(1, write_bytes))
|
||||
if b & 0b10000000 == 0:
|
||||
obj = b
|
||||
elif b & 0b11100000 == 0b11100000:
|
||||
obj = struct.unpack("b", chr(b))[0]
|
||||
elif b & 0b11100000 == 0b10100000:
|
||||
n = b & 0b00011111
|
||||
obj = struct.unpack("%ds" % n, self._fb_read(n, write_bytes))[0]
|
||||
typ = TYPE_RAW
|
||||
elif b & 0b11110000 == 0b10010000:
|
||||
n = b & 0b00001111
|
||||
typ = TYPE_ARRAY
|
||||
elif b & 0b11110000 == 0b10000000:
|
||||
n = b & 0b00001111
|
||||
typ = TYPE_MAP
|
||||
elif b == 0xc0:
|
||||
obj = None
|
||||
elif b == 0xc2:
|
||||
obj = False
|
||||
elif b == 0xc3:
|
||||
obj = True
|
||||
elif b == 0xca:
|
||||
obj = struct.unpack(">f", self._fb_read(4, write_bytes))[0]
|
||||
elif b == 0xcb:
|
||||
obj = struct.unpack(">d", self._fb_read(8, write_bytes))[0]
|
||||
elif b == 0xcc:
|
||||
obj = struct.unpack("B", self._fb_read(1, write_bytes))[0]
|
||||
elif b == 0xcd:
|
||||
obj = struct.unpack(">H", self._fb_read(2, write_bytes))[0]
|
||||
elif b == 0xce:
|
||||
obj = struct.unpack(">I", self._fb_read(4, write_bytes))[0]
|
||||
elif b == 0xcf:
|
||||
obj = struct.unpack(">Q", self._fb_read(8, write_bytes))[0]
|
||||
elif b == 0xd0:
|
||||
obj = struct.unpack("b", self._fb_read(1, write_bytes))[0]
|
||||
elif b == 0xd1:
|
||||
obj = struct.unpack(">h", self._fb_read(2, write_bytes))[0]
|
||||
elif b == 0xd2:
|
||||
obj = struct.unpack(">i", self._fb_read(4, write_bytes))[0]
|
||||
elif b == 0xd3:
|
||||
obj = struct.unpack(">q", self._fb_read(8, write_bytes))[0]
|
||||
elif b == 0xda:
|
||||
n = struct.unpack(">H", self._fb_read(2, write_bytes))[0]
|
||||
obj = struct.unpack("%ds" % n, self._fb_read(n, write_bytes))[0]
|
||||
typ = TYPE_RAW
|
||||
elif b == 0xdb:
|
||||
n = struct.unpack(">I", self._fb_read(4, write_bytes))[0]
|
||||
obj = struct.unpack("%ds" % n, self._fb_read(n, write_bytes))[0]
|
||||
typ = TYPE_RAW
|
||||
elif b == 0xdc:
|
||||
n = struct.unpack(">H", self._fb_read(2, write_bytes))[0]
|
||||
typ = TYPE_ARRAY
|
||||
elif b == 0xdd:
|
||||
n = struct.unpack(">I", self._fb_read(4, write_bytes))[0]
|
||||
typ = TYPE_ARRAY
|
||||
elif b == 0xde:
|
||||
n = struct.unpack(">H", self._fb_read(2, write_bytes))[0]
|
||||
typ = TYPE_MAP
|
||||
elif b == 0xdf:
|
||||
n = struct.unpack(">I", self._fb_read(4, write_bytes))[0]
|
||||
typ = TYPE_MAP
|
||||
else:
|
||||
raise UnpackValueError("Unknown header: 0x%x" % b)
|
||||
if execute == EX_READ_ARRAY_HEADER:
|
||||
if typ != TYPE_ARRAY:
|
||||
raise UnpackValueError("Expected array")
|
||||
return n
|
||||
if execute == EX_READ_MAP_HEADER:
|
||||
if typ != TYPE_MAP:
|
||||
raise UnpackValueError("Expected map")
|
||||
return n
|
||||
# TODO should we eliminate the recursion?
|
||||
if typ == TYPE_ARRAY:
|
||||
if execute == EX_SKIP:
|
||||
for i in xrange(n):
|
||||
# TODO check whether we need to call `list_hook`
|
||||
self._fb_unpack(EX_SKIP, write_bytes)
|
||||
return
|
||||
ret = []
|
||||
for i in xrange(n):
|
||||
ret.append(self._fb_unpack(EX_CONSTRUCT, write_bytes))
|
||||
if self.list_hook is not None:
|
||||
ret = self.list_hook(ret)
|
||||
# TODO is the interaction between `list_hook` and `use_list` ok?
|
||||
return ret if self.use_list else tuple(ret)
|
||||
if typ == TYPE_MAP:
|
||||
if execute == EX_SKIP:
|
||||
for i in xrange(n):
|
||||
# TODO check whether we need to call hooks
|
||||
self._fb_unpack(EX_SKIP, write_bytes)
|
||||
self._fb_unpack(EX_SKIP, write_bytes)
|
||||
return
|
||||
ret = []
|
||||
for i in xrange(n):
|
||||
ret.append((self._fb_unpack(EX_CONSTRUCT, write_bytes),
|
||||
self._fb_unpack(EX_CONSTRUCT, write_bytes)))
|
||||
if self.object_pairs_hook is not None:
|
||||
ret = self.object_pairs_hook(ret)
|
||||
else:
|
||||
ret = dict(ret)
|
||||
if self.object_hook is not None:
|
||||
ret = self.object_hook(ret)
|
||||
return ret
|
||||
if execute == EX_SKIP:
|
||||
return
|
||||
if typ == TYPE_RAW:
|
||||
if self.encoding is not None:
|
||||
obj = obj.decode(self.encoding, self.unicode_errors)
|
||||
return obj
|
||||
assert typ == TYPE_IMMEDIATE
|
||||
return obj
|
||||
|
||||
def skip(self, write_bytes=None):
|
||||
self._fb_unpack(EX_SKIP, write_bytes)
|
||||
self._fb_consume()
|
||||
|
||||
def unpack(self, write_bytes=None):
|
||||
ret = self._fb_unpack(EX_CONSTRUCT, write_bytes)
|
||||
self._fb_consume()
|
||||
return ret
|
||||
|
||||
def read_array_header(self, write_bytes=None):
|
||||
ret = self._fb_unpack(EX_READ_ARRAY_HEADER, write_bytes)
|
||||
self._fb_consume()
|
||||
return ret
|
||||
|
||||
def read_map_header(self, write_bytes=None):
|
||||
ret = self._fb_unpack(EX_READ_MAP_HEADER, write_bytes)
|
||||
self._fb_consume()
|
||||
return ret
|
||||
|
||||
class Packer(object):
|
||||
def __init__(self, default=None, encoding='utf-8', unicode_errors='strict',
|
||||
use_single_float=False, autoreset=True):
|
||||
self.use_float = use_single_float
|
||||
self.autoreset = autoreset
|
||||
self.encoding = encoding
|
||||
self.unicode_errors = unicode_errors
|
||||
self.buffer = StringIO.StringIO()
|
||||
if default is not None:
|
||||
if not callable(default):
|
||||
raise TypeError("default must be callable")
|
||||
self._default = default
|
||||
def _pack(self, obj, nest_limit=DEFAULT_RECURSE_LIMIT):
|
||||
if nest_limit < 0:
|
||||
raise PackValueError("recursion limit exceeded")
|
||||
if obj is None:
|
||||
return self.buffer.write(chr(0xc0))
|
||||
if isinstance(obj, bool):
|
||||
if obj:
|
||||
return self.buffer.write(chr(0xc3))
|
||||
return self.buffer.write(chr(0xc2))
|
||||
if isinstance(obj, int) or isinstance(obj, long):
|
||||
if 0 <= obj < 0x80:
|
||||
return self.buffer.write(struct.pack("B", obj))
|
||||
if -0x20 <= obj < 0:
|
||||
return self.buffer.write(struct.pack("b", obj))
|
||||
if 0x80 <= obj <= 0xff:
|
||||
return self.buffer.write(struct.pack("BB", 0xcc, obj))
|
||||
if -0x80 <= obj < 0:
|
||||
return self.buffer.write(struct.pack(">Bb", 0xd0, obj))
|
||||
if 0xff < obj <= 0xffff:
|
||||
return self.buffer.write(struct.pack(">BH", 0xcd, obj))
|
||||
if -0x8000 <= obj < -0x80:
|
||||
return self.buffer.write(struct.pack(">Bh", 0xd1, obj))
|
||||
if 0xffff < obj <= 0xffffffff:
|
||||
return self.buffer.write(struct.pack(">BI", 0xce, obj))
|
||||
if -0x80000000 <= obj < -0x8000:
|
||||
return self.buffer.write(struct.pack(">Bi", 0xd2, obj))
|
||||
if 0xffffffff < obj <= 0xffffffffffffffff:
|
||||
return self.buffer.write(struct.pack(">BQ", 0xcf, obj))
|
||||
if -0x8000000000000000 <= obj < -0x80000000:
|
||||
return self.buffer.write(struct.pack(">Bq", 0xd3, obj))
|
||||
raise PackValueError("Integer value out of range")
|
||||
if isinstance(obj, str) or isinstance(obj, unicode):
|
||||
if isinstance(obj, unicode):
|
||||
obj = obj.encode(self.encoding, self.unicode_errors)
|
||||
n = len(obj)
|
||||
if n <= 0x1f:
|
||||
return self.buffer.write(chr(0xa0 + n) + obj)
|
||||
if n <= 0xffff:
|
||||
return self.buffer.write(struct.pack(">BH%ds" % n,0xda, n, obj))
|
||||
if n <= 0xffffffff:
|
||||
return self.buffer.write(struct.pack(">BI%ds" % n,0xdb, n, obj))
|
||||
raise PackValueError("String is too large")
|
||||
if isinstance(obj, float):
|
||||
if self.use_float:
|
||||
return self.buffer.write(struct.pack(">Bf", 0xca, obj))
|
||||
return self.buffer.write(struct.pack(">Bd", 0xcb, obj))
|
||||
if isinstance(obj, list) or isinstance(obj, tuple):
|
||||
n = len(obj)
|
||||
self._fb_pack_array_header(n)
|
||||
for i in xrange(n):
|
||||
self._pack(obj[i], nest_limit - 1)
|
||||
return
|
||||
if isinstance(obj, dict):
|
||||
return self._fb_pack_map_pairs(len(obj), obj.iteritems(),
|
||||
nest_limit - 1)
|
||||
if self._default is not None:
|
||||
return self._pack(self._default(obj), nest_limit - 1)
|
||||
raise TypeError("Cannot serialize %r" % obj)
|
||||
def pack(self, obj):
|
||||
self._pack(obj)
|
||||
ret = self.buffer.getvalue()
|
||||
if self.autoreset:
|
||||
self.buffer = StringIO.StringIO()
|
||||
return ret
|
||||
def pack_map_pairs(self, pairs):
|
||||
self._fb_pack_map_pairs(len(pairs), pairs)
|
||||
ret = self.buffer.getvalue()
|
||||
if self.autoreset:
|
||||
self.buffer = StringIO.StringIO()
|
||||
return ret
|
||||
def pack_array_header(self, n):
|
||||
self._fb_pack_array_header(n)
|
||||
ret = self.buffer.getvalue()
|
||||
if self.autoreset:
|
||||
self.buffer = StringIO.StringIO()
|
||||
return ret
|
||||
def pack_map_header(self, n):
|
||||
self._fb_pack_map_header(n)
|
||||
ret = self.buffer.getvalue()
|
||||
if self.autoreset:
|
||||
self.buffer = StringIO.StringIO()
|
||||
return ret
|
||||
def _fb_pack_array_header(self, n):
|
||||
if n <= 0x0f:
|
||||
return self.buffer.write(chr(0x90 + n))
|
||||
if n <= 0xffff:
|
||||
return self.buffer.write(struct.pack(">BH", 0xdc, n))
|
||||
if n <= 0xffffffff:
|
||||
return self.buffer.write(struct.pack(">BI", 0xdd, n))
|
||||
raise PackValueError("Array is too large")
|
||||
def _fb_pack_map_header(self, n):
|
||||
if n <= 0x0f:
|
||||
return self.buffer.write(chr(0x80 + n))
|
||||
if n <= 0xffff:
|
||||
return self.buffer.write(struct.pack(">BH", 0xde, n))
|
||||
if n <= 0xffffffff:
|
||||
return self.buffer.write(struct.pack(">BI", 0xdf, n))
|
||||
raise PackValueError("Dict is too large")
|
||||
def _fb_pack_map_pairs(self, n, pairs, nest_limit=DEFAULT_RECURSE_LIMIT):
|
||||
self._fb_pack_map_header(n)
|
||||
for (k, v) in pairs:
|
||||
self._pack(k, nest_limit - 1)
|
||||
self._pack(v, nest_limit - 1)
|
||||
def bytes(self):
|
||||
return self.buffer.getvalue()
|
||||
def reset(self):
|
||||
self.buffer = StringIO.StringIO()
|
Loading…
Add table
Add a link
Reference in a new issue