mirror of
				https://github.com/python/cpython.git
				synced 2025-11-02 14:41:33 +00:00 
			
		
		
		
	Skip test_pair() and test_speech128() of test_zlib on s390x since they fail if zlib uses the s390x hardware accelerator.
		
			
				
	
	
		
			953 lines
		
	
	
	
		
			36 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			953 lines
		
	
	
	
		
			36 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
import unittest
 | 
						|
from test import support
 | 
						|
from test.support import import_helper
 | 
						|
import binascii
 | 
						|
import copy
 | 
						|
import os
 | 
						|
import pickle
 | 
						|
import random
 | 
						|
import sys
 | 
						|
from test.support import bigmemtest, _1G, _4G
 | 
						|
 | 
						|
 | 
						|
zlib = import_helper.import_module('zlib')
 | 
						|
 | 
						|
requires_Compress_copy = unittest.skipUnless(
 | 
						|
        hasattr(zlib.compressobj(), "copy"),
 | 
						|
        'requires Compress.copy()')
 | 
						|
requires_Decompress_copy = unittest.skipUnless(
 | 
						|
        hasattr(zlib.decompressobj(), "copy"),
 | 
						|
        'requires Decompress.copy()')
 | 
						|
 | 
						|
# bpo-46623: On s390x, when a hardware accelerator is used, using different
 | 
						|
# ways to compress data with zlib can produce different compressed data.
 | 
						|
# Simplified test_pair() code:
 | 
						|
#
 | 
						|
#   def func1(data):
 | 
						|
#       return zlib.compress(data)
 | 
						|
#
 | 
						|
#   def func2(data)
 | 
						|
#       co = zlib.compressobj()
 | 
						|
#       x1 = co.compress(data)
 | 
						|
#       x2 = co.flush()
 | 
						|
#       return x1 + x2
 | 
						|
#
 | 
						|
# On s390x if zlib uses a hardware accelerator, func1() creates a single
 | 
						|
# "final" compressed block whereas func2() produces 3 compressed blocks (the
 | 
						|
# last one is a final block). On other platforms with no accelerator, func1()
 | 
						|
# and func2() produce the same compressed data made of a single (final)
 | 
						|
# compressed block.
 | 
						|
#
 | 
						|
# Only the compressed data is different, the decompression returns the original
 | 
						|
# data:
 | 
						|
#
 | 
						|
#   zlib.decompress(func1(data)) == zlib.decompress(func2(data)) == data
 | 
						|
#
 | 
						|
# Make the assumption that s390x always has an accelerator to simplify the skip
 | 
						|
# condition. Windows doesn't have os.uname() but it doesn't support s390x.
 | 
						|
skip_on_s390x = unittest.skipIf(hasattr(os, 'uname') and os.uname().machine == 's390x',
 | 
						|
                                'skipped on s390x')
 | 
						|
 | 
						|
 | 
						|
class VersionTestCase(unittest.TestCase):
 | 
						|
 | 
						|
    def test_library_version(self):
 | 
						|
        # Test that the major version of the actual library in use matches the
 | 
						|
        # major version that we were compiled against. We can't guarantee that
 | 
						|
        # the minor versions will match (even on the machine on which the module
 | 
						|
        # was compiled), and the API is stable between minor versions, so
 | 
						|
        # testing only the major versions avoids spurious failures.
 | 
						|
        self.assertEqual(zlib.ZLIB_RUNTIME_VERSION[0], zlib.ZLIB_VERSION[0])
 | 
						|
 | 
						|
 | 
						|
class ChecksumTestCase(unittest.TestCase):
 | 
						|
    # checksum test cases
 | 
						|
    def test_crc32start(self):
 | 
						|
        self.assertEqual(zlib.crc32(b""), zlib.crc32(b"", 0))
 | 
						|
        self.assertTrue(zlib.crc32(b"abc", 0xffffffff))
 | 
						|
 | 
						|
    def test_crc32empty(self):
 | 
						|
        self.assertEqual(zlib.crc32(b"", 0), 0)
 | 
						|
        self.assertEqual(zlib.crc32(b"", 1), 1)
 | 
						|
        self.assertEqual(zlib.crc32(b"", 432), 432)
 | 
						|
 | 
						|
    def test_adler32start(self):
 | 
						|
        self.assertEqual(zlib.adler32(b""), zlib.adler32(b"", 1))
 | 
						|
        self.assertTrue(zlib.adler32(b"abc", 0xffffffff))
 | 
						|
 | 
						|
    def test_adler32empty(self):
 | 
						|
        self.assertEqual(zlib.adler32(b"", 0), 0)
 | 
						|
        self.assertEqual(zlib.adler32(b"", 1), 1)
 | 
						|
        self.assertEqual(zlib.adler32(b"", 432), 432)
 | 
						|
 | 
						|
    def test_penguins(self):
 | 
						|
        self.assertEqual(zlib.crc32(b"penguin", 0), 0x0e5c1a120)
 | 
						|
        self.assertEqual(zlib.crc32(b"penguin", 1), 0x43b6aa94)
 | 
						|
        self.assertEqual(zlib.adler32(b"penguin", 0), 0x0bcf02f6)
 | 
						|
        self.assertEqual(zlib.adler32(b"penguin", 1), 0x0bd602f7)
 | 
						|
 | 
						|
        self.assertEqual(zlib.crc32(b"penguin"), zlib.crc32(b"penguin", 0))
 | 
						|
        self.assertEqual(zlib.adler32(b"penguin"),zlib.adler32(b"penguin",1))
 | 
						|
 | 
						|
    def test_crc32_adler32_unsigned(self):
 | 
						|
        foo = b'abcdefghijklmnop'
 | 
						|
        # explicitly test signed behavior
 | 
						|
        self.assertEqual(zlib.crc32(foo), 2486878355)
 | 
						|
        self.assertEqual(zlib.crc32(b'spam'), 1138425661)
 | 
						|
        self.assertEqual(zlib.adler32(foo+foo), 3573550353)
 | 
						|
        self.assertEqual(zlib.adler32(b'spam'), 72286642)
 | 
						|
 | 
						|
    def test_same_as_binascii_crc32(self):
 | 
						|
        foo = b'abcdefghijklmnop'
 | 
						|
        crc = 2486878355
 | 
						|
        self.assertEqual(binascii.crc32(foo), crc)
 | 
						|
        self.assertEqual(zlib.crc32(foo), crc)
 | 
						|
        self.assertEqual(binascii.crc32(b'spam'), zlib.crc32(b'spam'))
 | 
						|
 | 
						|
 | 
						|
# Issue #10276 - check that inputs >=4 GiB are handled correctly.
 | 
						|
class ChecksumBigBufferTestCase(unittest.TestCase):
 | 
						|
 | 
						|
    @bigmemtest(size=_4G + 4, memuse=1, dry_run=False)
 | 
						|
    def test_big_buffer(self, size):
 | 
						|
        data = b"nyan" * (_1G + 1)
 | 
						|
        self.assertEqual(zlib.crc32(data), 1044521549)
 | 
						|
        self.assertEqual(zlib.adler32(data), 2256789997)
 | 
						|
 | 
						|
 | 
						|
class ExceptionTestCase(unittest.TestCase):
 | 
						|
    # make sure we generate some expected errors
 | 
						|
    def test_badlevel(self):
 | 
						|
        # specifying compression level out of range causes an error
 | 
						|
        # (but -1 is Z_DEFAULT_COMPRESSION and apparently the zlib
 | 
						|
        # accepts 0 too)
 | 
						|
        self.assertRaises(zlib.error, zlib.compress, b'ERROR', 10)
 | 
						|
 | 
						|
    def test_badargs(self):
 | 
						|
        self.assertRaises(TypeError, zlib.adler32)
 | 
						|
        self.assertRaises(TypeError, zlib.crc32)
 | 
						|
        self.assertRaises(TypeError, zlib.compress)
 | 
						|
        self.assertRaises(TypeError, zlib.decompress)
 | 
						|
        for arg in (42, None, '', 'abc', (), []):
 | 
						|
            self.assertRaises(TypeError, zlib.adler32, arg)
 | 
						|
            self.assertRaises(TypeError, zlib.crc32, arg)
 | 
						|
            self.assertRaises(TypeError, zlib.compress, arg)
 | 
						|
            self.assertRaises(TypeError, zlib.decompress, arg)
 | 
						|
 | 
						|
    def test_badcompressobj(self):
 | 
						|
        # verify failure on building compress object with bad params
 | 
						|
        self.assertRaises(ValueError, zlib.compressobj, 1, zlib.DEFLATED, 0)
 | 
						|
        # specifying total bits too large causes an error
 | 
						|
        self.assertRaises(ValueError,
 | 
						|
                zlib.compressobj, 1, zlib.DEFLATED, zlib.MAX_WBITS + 1)
 | 
						|
 | 
						|
    def test_baddecompressobj(self):
 | 
						|
        # verify failure on building decompress object with bad params
 | 
						|
        self.assertRaises(ValueError, zlib.decompressobj, -1)
 | 
						|
 | 
						|
    def test_decompressobj_badflush(self):
 | 
						|
        # verify failure on calling decompressobj.flush with bad params
 | 
						|
        self.assertRaises(ValueError, zlib.decompressobj().flush, 0)
 | 
						|
        self.assertRaises(ValueError, zlib.decompressobj().flush, -1)
 | 
						|
 | 
						|
    @support.cpython_only
 | 
						|
    def test_overflow(self):
 | 
						|
        with self.assertRaisesRegex(OverflowError, 'int too large'):
 | 
						|
            zlib.decompress(b'', 15, sys.maxsize + 1)
 | 
						|
        with self.assertRaisesRegex(OverflowError, 'int too large'):
 | 
						|
            zlib.decompressobj().decompress(b'', sys.maxsize + 1)
 | 
						|
        with self.assertRaisesRegex(OverflowError, 'int too large'):
 | 
						|
            zlib.decompressobj().flush(sys.maxsize + 1)
 | 
						|
 | 
						|
    @support.cpython_only
 | 
						|
    def test_disallow_instantiation(self):
 | 
						|
        # Ensure that the type disallows instantiation (bpo-43916)
 | 
						|
        support.check_disallow_instantiation(self, type(zlib.compressobj()))
 | 
						|
        support.check_disallow_instantiation(self, type(zlib.decompressobj()))
 | 
						|
 | 
						|
 | 
						|
class BaseCompressTestCase(object):
 | 
						|
    def check_big_compress_buffer(self, size, compress_func):
 | 
						|
        _1M = 1024 * 1024
 | 
						|
        # Generate 10 MiB worth of random, and expand it by repeating it.
 | 
						|
        # The assumption is that zlib's memory is not big enough to exploit
 | 
						|
        # such spread out redundancy.
 | 
						|
        data = random.randbytes(_1M * 10)
 | 
						|
        data = data * (size // len(data) + 1)
 | 
						|
        try:
 | 
						|
            compress_func(data)
 | 
						|
        finally:
 | 
						|
            # Release memory
 | 
						|
            data = None
 | 
						|
 | 
						|
    def check_big_decompress_buffer(self, size, decompress_func):
 | 
						|
        data = b'x' * size
 | 
						|
        try:
 | 
						|
            compressed = zlib.compress(data, 1)
 | 
						|
        finally:
 | 
						|
            # Release memory
 | 
						|
            data = None
 | 
						|
        data = decompress_func(compressed)
 | 
						|
        # Sanity check
 | 
						|
        try:
 | 
						|
            self.assertEqual(len(data), size)
 | 
						|
            self.assertEqual(len(data.strip(b'x')), 0)
 | 
						|
        finally:
 | 
						|
            data = None
 | 
						|
 | 
						|
 | 
						|
class CompressTestCase(BaseCompressTestCase, unittest.TestCase):
 | 
						|
    # Test compression in one go (whole message compression)
 | 
						|
    def test_speech(self):
 | 
						|
        x = zlib.compress(HAMLET_SCENE)
 | 
						|
        self.assertEqual(zlib.decompress(x), HAMLET_SCENE)
 | 
						|
 | 
						|
    def test_keywords(self):
 | 
						|
        x = zlib.compress(HAMLET_SCENE, level=3)
 | 
						|
        self.assertEqual(zlib.decompress(x), HAMLET_SCENE)
 | 
						|
        with self.assertRaises(TypeError):
 | 
						|
            zlib.compress(data=HAMLET_SCENE, level=3)
 | 
						|
        self.assertEqual(zlib.decompress(x,
 | 
						|
                                         wbits=zlib.MAX_WBITS,
 | 
						|
                                         bufsize=zlib.DEF_BUF_SIZE),
 | 
						|
                         HAMLET_SCENE)
 | 
						|
 | 
						|
    @skip_on_s390x
 | 
						|
    def test_speech128(self):
 | 
						|
        # compress more data
 | 
						|
        data = HAMLET_SCENE * 128
 | 
						|
        x = zlib.compress(data)
 | 
						|
        self.assertEqual(zlib.compress(bytearray(data)), x)
 | 
						|
        for ob in x, bytearray(x):
 | 
						|
            self.assertEqual(zlib.decompress(ob), data)
 | 
						|
 | 
						|
    def test_incomplete_stream(self):
 | 
						|
        # A useful error message is given
 | 
						|
        x = zlib.compress(HAMLET_SCENE)
 | 
						|
        self.assertRaisesRegex(zlib.error,
 | 
						|
            "Error -5 while decompressing data: incomplete or truncated stream",
 | 
						|
            zlib.decompress, x[:-1])
 | 
						|
 | 
						|
    # Memory use of the following functions takes into account overallocation
 | 
						|
 | 
						|
    @bigmemtest(size=_1G + 1024 * 1024, memuse=3)
 | 
						|
    def test_big_compress_buffer(self, size):
 | 
						|
        compress = lambda s: zlib.compress(s, 1)
 | 
						|
        self.check_big_compress_buffer(size, compress)
 | 
						|
 | 
						|
    @bigmemtest(size=_1G + 1024 * 1024, memuse=2)
 | 
						|
    def test_big_decompress_buffer(self, size):
 | 
						|
        self.check_big_decompress_buffer(size, zlib.decompress)
 | 
						|
 | 
						|
    @bigmemtest(size=_4G, memuse=1)
 | 
						|
    def test_large_bufsize(self, size):
 | 
						|
        # Test decompress(bufsize) parameter greater than the internal limit
 | 
						|
        data = HAMLET_SCENE * 10
 | 
						|
        compressed = zlib.compress(data, 1)
 | 
						|
        self.assertEqual(zlib.decompress(compressed, 15, size), data)
 | 
						|
 | 
						|
    def test_custom_bufsize(self):
 | 
						|
        data = HAMLET_SCENE * 10
 | 
						|
        compressed = zlib.compress(data, 1)
 | 
						|
        self.assertEqual(zlib.decompress(compressed, 15, CustomInt()), data)
 | 
						|
 | 
						|
    @unittest.skipUnless(sys.maxsize > 2**32, 'requires 64bit platform')
 | 
						|
    @bigmemtest(size=_4G + 100, memuse=4)
 | 
						|
    def test_64bit_compress(self, size):
 | 
						|
        data = b'x' * size
 | 
						|
        try:
 | 
						|
            comp = zlib.compress(data, 0)
 | 
						|
            self.assertEqual(zlib.decompress(comp), data)
 | 
						|
        finally:
 | 
						|
            comp = data = None
 | 
						|
 | 
						|
 | 
						|
class CompressObjectTestCase(BaseCompressTestCase, unittest.TestCase):
 | 
						|
    # Test compression object
 | 
						|
    @skip_on_s390x
 | 
						|
    def test_pair(self):
 | 
						|
        # straightforward compress/decompress objects
 | 
						|
        datasrc = HAMLET_SCENE * 128
 | 
						|
        datazip = zlib.compress(datasrc)
 | 
						|
        # should compress both bytes and bytearray data
 | 
						|
        for data in (datasrc, bytearray(datasrc)):
 | 
						|
            co = zlib.compressobj()
 | 
						|
            x1 = co.compress(data)
 | 
						|
            x2 = co.flush()
 | 
						|
            self.assertRaises(zlib.error, co.flush) # second flush should not work
 | 
						|
            self.assertEqual(x1 + x2, datazip)
 | 
						|
        for v1, v2 in ((x1, x2), (bytearray(x1), bytearray(x2))):
 | 
						|
            dco = zlib.decompressobj()
 | 
						|
            y1 = dco.decompress(v1 + v2)
 | 
						|
            y2 = dco.flush()
 | 
						|
            self.assertEqual(data, y1 + y2)
 | 
						|
            self.assertIsInstance(dco.unconsumed_tail, bytes)
 | 
						|
            self.assertIsInstance(dco.unused_data, bytes)
 | 
						|
 | 
						|
    def test_keywords(self):
 | 
						|
        level = 2
 | 
						|
        method = zlib.DEFLATED
 | 
						|
        wbits = -12
 | 
						|
        memLevel = 9
 | 
						|
        strategy = zlib.Z_FILTERED
 | 
						|
        co = zlib.compressobj(level=level,
 | 
						|
                              method=method,
 | 
						|
                              wbits=wbits,
 | 
						|
                              memLevel=memLevel,
 | 
						|
                              strategy=strategy,
 | 
						|
                              zdict=b"")
 | 
						|
        do = zlib.decompressobj(wbits=wbits, zdict=b"")
 | 
						|
        with self.assertRaises(TypeError):
 | 
						|
            co.compress(data=HAMLET_SCENE)
 | 
						|
        with self.assertRaises(TypeError):
 | 
						|
            do.decompress(data=zlib.compress(HAMLET_SCENE))
 | 
						|
        x = co.compress(HAMLET_SCENE) + co.flush()
 | 
						|
        y = do.decompress(x, max_length=len(HAMLET_SCENE)) + do.flush()
 | 
						|
        self.assertEqual(HAMLET_SCENE, y)
 | 
						|
 | 
						|
    def test_compressoptions(self):
 | 
						|
        # specify lots of options to compressobj()
 | 
						|
        level = 2
 | 
						|
        method = zlib.DEFLATED
 | 
						|
        wbits = -12
 | 
						|
        memLevel = 9
 | 
						|
        strategy = zlib.Z_FILTERED
 | 
						|
        co = zlib.compressobj(level, method, wbits, memLevel, strategy)
 | 
						|
        x1 = co.compress(HAMLET_SCENE)
 | 
						|
        x2 = co.flush()
 | 
						|
        dco = zlib.decompressobj(wbits)
 | 
						|
        y1 = dco.decompress(x1 + x2)
 | 
						|
        y2 = dco.flush()
 | 
						|
        self.assertEqual(HAMLET_SCENE, y1 + y2)
 | 
						|
 | 
						|
    def test_compressincremental(self):
 | 
						|
        # compress object in steps, decompress object as one-shot
 | 
						|
        data = HAMLET_SCENE * 128
 | 
						|
        co = zlib.compressobj()
 | 
						|
        bufs = []
 | 
						|
        for i in range(0, len(data), 256):
 | 
						|
            bufs.append(co.compress(data[i:i+256]))
 | 
						|
        bufs.append(co.flush())
 | 
						|
        combuf = b''.join(bufs)
 | 
						|
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        y1 = dco.decompress(b''.join(bufs))
 | 
						|
        y2 = dco.flush()
 | 
						|
        self.assertEqual(data, y1 + y2)
 | 
						|
 | 
						|
    def test_decompinc(self, flush=False, source=None, cx=256, dcx=64):
 | 
						|
        # compress object in steps, decompress object in steps
 | 
						|
        source = source or HAMLET_SCENE
 | 
						|
        data = source * 128
 | 
						|
        co = zlib.compressobj()
 | 
						|
        bufs = []
 | 
						|
        for i in range(0, len(data), cx):
 | 
						|
            bufs.append(co.compress(data[i:i+cx]))
 | 
						|
        bufs.append(co.flush())
 | 
						|
        combuf = b''.join(bufs)
 | 
						|
 | 
						|
        decombuf = zlib.decompress(combuf)
 | 
						|
        # Test type of return value
 | 
						|
        self.assertIsInstance(decombuf, bytes)
 | 
						|
 | 
						|
        self.assertEqual(data, decombuf)
 | 
						|
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        bufs = []
 | 
						|
        for i in range(0, len(combuf), dcx):
 | 
						|
            bufs.append(dco.decompress(combuf[i:i+dcx]))
 | 
						|
            self.assertEqual(b'', dco.unconsumed_tail, ########
 | 
						|
                             "(A) uct should be b'': not %d long" %
 | 
						|
                                       len(dco.unconsumed_tail))
 | 
						|
            self.assertEqual(b'', dco.unused_data)
 | 
						|
        if flush:
 | 
						|
            bufs.append(dco.flush())
 | 
						|
        else:
 | 
						|
            while True:
 | 
						|
                chunk = dco.decompress(b'')
 | 
						|
                if chunk:
 | 
						|
                    bufs.append(chunk)
 | 
						|
                else:
 | 
						|
                    break
 | 
						|
        self.assertEqual(b'', dco.unconsumed_tail, ########
 | 
						|
                         "(B) uct should be b'': not %d long" %
 | 
						|
                                       len(dco.unconsumed_tail))
 | 
						|
        self.assertEqual(b'', dco.unused_data)
 | 
						|
        self.assertEqual(data, b''.join(bufs))
 | 
						|
        # Failure means: "decompressobj with init options failed"
 | 
						|
 | 
						|
    def test_decompincflush(self):
 | 
						|
        self.test_decompinc(flush=True)
 | 
						|
 | 
						|
    def test_decompimax(self, source=None, cx=256, dcx=64):
 | 
						|
        # compress in steps, decompress in length-restricted steps
 | 
						|
        source = source or HAMLET_SCENE
 | 
						|
        # Check a decompression object with max_length specified
 | 
						|
        data = source * 128
 | 
						|
        co = zlib.compressobj()
 | 
						|
        bufs = []
 | 
						|
        for i in range(0, len(data), cx):
 | 
						|
            bufs.append(co.compress(data[i:i+cx]))
 | 
						|
        bufs.append(co.flush())
 | 
						|
        combuf = b''.join(bufs)
 | 
						|
        self.assertEqual(data, zlib.decompress(combuf),
 | 
						|
                         'compressed data failure')
 | 
						|
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        bufs = []
 | 
						|
        cb = combuf
 | 
						|
        while cb:
 | 
						|
            #max_length = 1 + len(cb)//10
 | 
						|
            chunk = dco.decompress(cb, dcx)
 | 
						|
            self.assertFalse(len(chunk) > dcx,
 | 
						|
                    'chunk too big (%d>%d)' % (len(chunk), dcx))
 | 
						|
            bufs.append(chunk)
 | 
						|
            cb = dco.unconsumed_tail
 | 
						|
        bufs.append(dco.flush())
 | 
						|
        self.assertEqual(data, b''.join(bufs), 'Wrong data retrieved')
 | 
						|
 | 
						|
    def test_decompressmaxlen(self, flush=False):
 | 
						|
        # Check a decompression object with max_length specified
 | 
						|
        data = HAMLET_SCENE * 128
 | 
						|
        co = zlib.compressobj()
 | 
						|
        bufs = []
 | 
						|
        for i in range(0, len(data), 256):
 | 
						|
            bufs.append(co.compress(data[i:i+256]))
 | 
						|
        bufs.append(co.flush())
 | 
						|
        combuf = b''.join(bufs)
 | 
						|
        self.assertEqual(data, zlib.decompress(combuf),
 | 
						|
                         'compressed data failure')
 | 
						|
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        bufs = []
 | 
						|
        cb = combuf
 | 
						|
        while cb:
 | 
						|
            max_length = 1 + len(cb)//10
 | 
						|
            chunk = dco.decompress(cb, max_length)
 | 
						|
            self.assertFalse(len(chunk) > max_length,
 | 
						|
                        'chunk too big (%d>%d)' % (len(chunk),max_length))
 | 
						|
            bufs.append(chunk)
 | 
						|
            cb = dco.unconsumed_tail
 | 
						|
        if flush:
 | 
						|
            bufs.append(dco.flush())
 | 
						|
        else:
 | 
						|
            while chunk:
 | 
						|
                chunk = dco.decompress(b'', max_length)
 | 
						|
                self.assertFalse(len(chunk) > max_length,
 | 
						|
                            'chunk too big (%d>%d)' % (len(chunk),max_length))
 | 
						|
                bufs.append(chunk)
 | 
						|
        self.assertEqual(data, b''.join(bufs), 'Wrong data retrieved')
 | 
						|
 | 
						|
    def test_decompressmaxlenflush(self):
 | 
						|
        self.test_decompressmaxlen(flush=True)
 | 
						|
 | 
						|
    def test_maxlenmisc(self):
 | 
						|
        # Misc tests of max_length
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        self.assertRaises(ValueError, dco.decompress, b"", -1)
 | 
						|
        self.assertEqual(b'', dco.unconsumed_tail)
 | 
						|
 | 
						|
    def test_maxlen_large(self):
 | 
						|
        # Sizes up to sys.maxsize should be accepted, although zlib is
 | 
						|
        # internally limited to expressing sizes with unsigned int
 | 
						|
        data = HAMLET_SCENE * 10
 | 
						|
        self.assertGreater(len(data), zlib.DEF_BUF_SIZE)
 | 
						|
        compressed = zlib.compress(data, 1)
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        self.assertEqual(dco.decompress(compressed, sys.maxsize), data)
 | 
						|
 | 
						|
    def test_maxlen_custom(self):
 | 
						|
        data = HAMLET_SCENE * 10
 | 
						|
        compressed = zlib.compress(data, 1)
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        self.assertEqual(dco.decompress(compressed, CustomInt()), data[:100])
 | 
						|
 | 
						|
    def test_clear_unconsumed_tail(self):
 | 
						|
        # Issue #12050: calling decompress() without providing max_length
 | 
						|
        # should clear the unconsumed_tail attribute.
 | 
						|
        cdata = b"x\x9cKLJ\x06\x00\x02M\x01"    # "abc"
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        ddata = dco.decompress(cdata, 1)
 | 
						|
        ddata += dco.decompress(dco.unconsumed_tail)
 | 
						|
        self.assertEqual(dco.unconsumed_tail, b"")
 | 
						|
 | 
						|
    def test_flushes(self):
 | 
						|
        # Test flush() with the various options, using all the
 | 
						|
        # different levels in order to provide more variations.
 | 
						|
        sync_opt = ['Z_NO_FLUSH', 'Z_SYNC_FLUSH', 'Z_FULL_FLUSH',
 | 
						|
                    'Z_PARTIAL_FLUSH']
 | 
						|
 | 
						|
        ver = tuple(int(v) for v in zlib.ZLIB_RUNTIME_VERSION.split('.'))
 | 
						|
        # Z_BLOCK has a known failure prior to 1.2.5.3
 | 
						|
        if ver >= (1, 2, 5, 3):
 | 
						|
            sync_opt.append('Z_BLOCK')
 | 
						|
 | 
						|
        sync_opt = [getattr(zlib, opt) for opt in sync_opt
 | 
						|
                    if hasattr(zlib, opt)]
 | 
						|
        data = HAMLET_SCENE * 8
 | 
						|
 | 
						|
        for sync in sync_opt:
 | 
						|
            for level in range(10):
 | 
						|
                try:
 | 
						|
                    obj = zlib.compressobj( level )
 | 
						|
                    a = obj.compress( data[:3000] )
 | 
						|
                    b = obj.flush( sync )
 | 
						|
                    c = obj.compress( data[3000:] )
 | 
						|
                    d = obj.flush()
 | 
						|
                except:
 | 
						|
                    print("Error for flush mode={}, level={}"
 | 
						|
                          .format(sync, level))
 | 
						|
                    raise
 | 
						|
                self.assertEqual(zlib.decompress(b''.join([a,b,c,d])),
 | 
						|
                                 data, ("Decompress failed: flush "
 | 
						|
                                        "mode=%i, level=%i") % (sync, level))
 | 
						|
                del obj
 | 
						|
 | 
						|
    @unittest.skipUnless(hasattr(zlib, 'Z_SYNC_FLUSH'),
 | 
						|
                         'requires zlib.Z_SYNC_FLUSH')
 | 
						|
    def test_odd_flush(self):
 | 
						|
        # Test for odd flushing bugs noted in 2.0, and hopefully fixed in 2.1
 | 
						|
        import random
 | 
						|
        # Testing on 17K of "random" data
 | 
						|
 | 
						|
        # Create compressor and decompressor objects
 | 
						|
        co = zlib.compressobj(zlib.Z_BEST_COMPRESSION)
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
 | 
						|
        # Try 17K of data
 | 
						|
        # generate random data stream
 | 
						|
        try:
 | 
						|
            # In 2.3 and later, WichmannHill is the RNG of the bug report
 | 
						|
            gen = random.WichmannHill()
 | 
						|
        except AttributeError:
 | 
						|
            try:
 | 
						|
                # 2.2 called it Random
 | 
						|
                gen = random.Random()
 | 
						|
            except AttributeError:
 | 
						|
                # others might simply have a single RNG
 | 
						|
                gen = random
 | 
						|
        gen.seed(1)
 | 
						|
        data = gen.randbytes(17 * 1024)
 | 
						|
 | 
						|
        # compress, sync-flush, and decompress
 | 
						|
        first = co.compress(data)
 | 
						|
        second = co.flush(zlib.Z_SYNC_FLUSH)
 | 
						|
        expanded = dco.decompress(first + second)
 | 
						|
 | 
						|
        # if decompressed data is different from the input data, choke.
 | 
						|
        self.assertEqual(expanded, data, "17K random source doesn't match")
 | 
						|
 | 
						|
    def test_empty_flush(self):
 | 
						|
        # Test that calling .flush() on unused objects works.
 | 
						|
        # (Bug #1083110 -- calling .flush() on decompress objects
 | 
						|
        # caused a core dump.)
 | 
						|
 | 
						|
        co = zlib.compressobj(zlib.Z_BEST_COMPRESSION)
 | 
						|
        self.assertTrue(co.flush())  # Returns a zlib header
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        self.assertEqual(dco.flush(), b"") # Returns nothing
 | 
						|
 | 
						|
    def test_dictionary(self):
 | 
						|
        h = HAMLET_SCENE
 | 
						|
        # Build a simulated dictionary out of the words in HAMLET.
 | 
						|
        words = h.split()
 | 
						|
        random.shuffle(words)
 | 
						|
        zdict = b''.join(words)
 | 
						|
        # Use it to compress HAMLET.
 | 
						|
        co = zlib.compressobj(zdict=zdict)
 | 
						|
        cd = co.compress(h) + co.flush()
 | 
						|
        # Verify that it will decompress with the dictionary.
 | 
						|
        dco = zlib.decompressobj(zdict=zdict)
 | 
						|
        self.assertEqual(dco.decompress(cd) + dco.flush(), h)
 | 
						|
        # Verify that it fails when not given the dictionary.
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        self.assertRaises(zlib.error, dco.decompress, cd)
 | 
						|
 | 
						|
    def test_dictionary_streaming(self):
 | 
						|
        # This simulates the reuse of a compressor object for compressing
 | 
						|
        # several separate data streams.
 | 
						|
        co = zlib.compressobj(zdict=HAMLET_SCENE)
 | 
						|
        do = zlib.decompressobj(zdict=HAMLET_SCENE)
 | 
						|
        piece = HAMLET_SCENE[1000:1500]
 | 
						|
        d0 = co.compress(piece) + co.flush(zlib.Z_SYNC_FLUSH)
 | 
						|
        d1 = co.compress(piece[100:]) + co.flush(zlib.Z_SYNC_FLUSH)
 | 
						|
        d2 = co.compress(piece[:-100]) + co.flush(zlib.Z_SYNC_FLUSH)
 | 
						|
        self.assertEqual(do.decompress(d0), piece)
 | 
						|
        self.assertEqual(do.decompress(d1), piece[100:])
 | 
						|
        self.assertEqual(do.decompress(d2), piece[:-100])
 | 
						|
 | 
						|
    def test_decompress_incomplete_stream(self):
 | 
						|
        # This is 'foo', deflated
 | 
						|
        x = b'x\x9cK\xcb\xcf\x07\x00\x02\x82\x01E'
 | 
						|
        # For the record
 | 
						|
        self.assertEqual(zlib.decompress(x), b'foo')
 | 
						|
        self.assertRaises(zlib.error, zlib.decompress, x[:-5])
 | 
						|
        # Omitting the stream end works with decompressor objects
 | 
						|
        # (see issue #8672).
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        y = dco.decompress(x[:-5])
 | 
						|
        y += dco.flush()
 | 
						|
        self.assertEqual(y, b'foo')
 | 
						|
 | 
						|
    def test_decompress_eof(self):
 | 
						|
        x = b'x\x9cK\xcb\xcf\x07\x00\x02\x82\x01E'  # 'foo'
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        self.assertFalse(dco.eof)
 | 
						|
        dco.decompress(x[:-5])
 | 
						|
        self.assertFalse(dco.eof)
 | 
						|
        dco.decompress(x[-5:])
 | 
						|
        self.assertTrue(dco.eof)
 | 
						|
        dco.flush()
 | 
						|
        self.assertTrue(dco.eof)
 | 
						|
 | 
						|
    def test_decompress_eof_incomplete_stream(self):
 | 
						|
        x = b'x\x9cK\xcb\xcf\x07\x00\x02\x82\x01E'  # 'foo'
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        self.assertFalse(dco.eof)
 | 
						|
        dco.decompress(x[:-5])
 | 
						|
        self.assertFalse(dco.eof)
 | 
						|
        dco.flush()
 | 
						|
        self.assertFalse(dco.eof)
 | 
						|
 | 
						|
    def test_decompress_unused_data(self):
 | 
						|
        # Repeated calls to decompress() after EOF should accumulate data in
 | 
						|
        # dco.unused_data, instead of just storing the arg to the last call.
 | 
						|
        source = b'abcdefghijklmnopqrstuvwxyz'
 | 
						|
        remainder = b'0123456789'
 | 
						|
        y = zlib.compress(source)
 | 
						|
        x = y + remainder
 | 
						|
        for maxlen in 0, 1000:
 | 
						|
            for step in 1, 2, len(y), len(x):
 | 
						|
                dco = zlib.decompressobj()
 | 
						|
                data = b''
 | 
						|
                for i in range(0, len(x), step):
 | 
						|
                    if i < len(y):
 | 
						|
                        self.assertEqual(dco.unused_data, b'')
 | 
						|
                    if maxlen == 0:
 | 
						|
                        data += dco.decompress(x[i : i + step])
 | 
						|
                        self.assertEqual(dco.unconsumed_tail, b'')
 | 
						|
                    else:
 | 
						|
                        data += dco.decompress(
 | 
						|
                                dco.unconsumed_tail + x[i : i + step], maxlen)
 | 
						|
                data += dco.flush()
 | 
						|
                self.assertTrue(dco.eof)
 | 
						|
                self.assertEqual(data, source)
 | 
						|
                self.assertEqual(dco.unconsumed_tail, b'')
 | 
						|
                self.assertEqual(dco.unused_data, remainder)
 | 
						|
 | 
						|
    # issue27164
 | 
						|
    def test_decompress_raw_with_dictionary(self):
 | 
						|
        zdict = b'abcdefghijklmnopqrstuvwxyz'
 | 
						|
        co = zlib.compressobj(wbits=-zlib.MAX_WBITS, zdict=zdict)
 | 
						|
        comp = co.compress(zdict) + co.flush()
 | 
						|
        dco = zlib.decompressobj(wbits=-zlib.MAX_WBITS, zdict=zdict)
 | 
						|
        uncomp = dco.decompress(comp) + dco.flush()
 | 
						|
        self.assertEqual(zdict, uncomp)
 | 
						|
 | 
						|
    def test_flush_with_freed_input(self):
 | 
						|
        # Issue #16411: decompressor accesses input to last decompress() call
 | 
						|
        # in flush(), even if this object has been freed in the meanwhile.
 | 
						|
        input1 = b'abcdefghijklmnopqrstuvwxyz'
 | 
						|
        input2 = b'QWERTYUIOPASDFGHJKLZXCVBNM'
 | 
						|
        data = zlib.compress(input1)
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        dco.decompress(data, 1)
 | 
						|
        del data
 | 
						|
        data = zlib.compress(input2)
 | 
						|
        self.assertEqual(dco.flush(), input1[1:])
 | 
						|
 | 
						|
    @bigmemtest(size=_4G, memuse=1)
 | 
						|
    def test_flush_large_length(self, size):
 | 
						|
        # Test flush(length) parameter greater than internal limit UINT_MAX
 | 
						|
        input = HAMLET_SCENE * 10
 | 
						|
        data = zlib.compress(input, 1)
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        dco.decompress(data, 1)
 | 
						|
        self.assertEqual(dco.flush(size), input[1:])
 | 
						|
 | 
						|
    def test_flush_custom_length(self):
 | 
						|
        input = HAMLET_SCENE * 10
 | 
						|
        data = zlib.compress(input, 1)
 | 
						|
        dco = zlib.decompressobj()
 | 
						|
        dco.decompress(data, 1)
 | 
						|
        self.assertEqual(dco.flush(CustomInt()), input[1:])
 | 
						|
 | 
						|
    @requires_Compress_copy
 | 
						|
    def test_compresscopy(self):
 | 
						|
        # Test copying a compression object
 | 
						|
        data0 = HAMLET_SCENE
 | 
						|
        data1 = bytes(str(HAMLET_SCENE, "ascii").swapcase(), "ascii")
 | 
						|
        for func in lambda c: c.copy(), copy.copy, copy.deepcopy:
 | 
						|
            c0 = zlib.compressobj(zlib.Z_BEST_COMPRESSION)
 | 
						|
            bufs0 = []
 | 
						|
            bufs0.append(c0.compress(data0))
 | 
						|
 | 
						|
            c1 = func(c0)
 | 
						|
            bufs1 = bufs0[:]
 | 
						|
 | 
						|
            bufs0.append(c0.compress(data0))
 | 
						|
            bufs0.append(c0.flush())
 | 
						|
            s0 = b''.join(bufs0)
 | 
						|
 | 
						|
            bufs1.append(c1.compress(data1))
 | 
						|
            bufs1.append(c1.flush())
 | 
						|
            s1 = b''.join(bufs1)
 | 
						|
 | 
						|
            self.assertEqual(zlib.decompress(s0),data0+data0)
 | 
						|
            self.assertEqual(zlib.decompress(s1),data0+data1)
 | 
						|
 | 
						|
    @requires_Compress_copy
 | 
						|
    def test_badcompresscopy(self):
 | 
						|
        # Test copying a compression object in an inconsistent state
 | 
						|
        c = zlib.compressobj()
 | 
						|
        c.compress(HAMLET_SCENE)
 | 
						|
        c.flush()
 | 
						|
        self.assertRaises(ValueError, c.copy)
 | 
						|
        self.assertRaises(ValueError, copy.copy, c)
 | 
						|
        self.assertRaises(ValueError, copy.deepcopy, c)
 | 
						|
 | 
						|
    @requires_Decompress_copy
 | 
						|
    def test_decompresscopy(self):
 | 
						|
        # Test copying a decompression object
 | 
						|
        data = HAMLET_SCENE
 | 
						|
        comp = zlib.compress(data)
 | 
						|
        # Test type of return value
 | 
						|
        self.assertIsInstance(comp, bytes)
 | 
						|
 | 
						|
        for func in lambda c: c.copy(), copy.copy, copy.deepcopy:
 | 
						|
            d0 = zlib.decompressobj()
 | 
						|
            bufs0 = []
 | 
						|
            bufs0.append(d0.decompress(comp[:32]))
 | 
						|
 | 
						|
            d1 = func(d0)
 | 
						|
            bufs1 = bufs0[:]
 | 
						|
 | 
						|
            bufs0.append(d0.decompress(comp[32:]))
 | 
						|
            s0 = b''.join(bufs0)
 | 
						|
 | 
						|
            bufs1.append(d1.decompress(comp[32:]))
 | 
						|
            s1 = b''.join(bufs1)
 | 
						|
 | 
						|
            self.assertEqual(s0,s1)
 | 
						|
            self.assertEqual(s0,data)
 | 
						|
 | 
						|
    @requires_Decompress_copy
 | 
						|
    def test_baddecompresscopy(self):
 | 
						|
        # Test copying a compression object in an inconsistent state
 | 
						|
        data = zlib.compress(HAMLET_SCENE)
 | 
						|
        d = zlib.decompressobj()
 | 
						|
        d.decompress(data)
 | 
						|
        d.flush()
 | 
						|
        self.assertRaises(ValueError, d.copy)
 | 
						|
        self.assertRaises(ValueError, copy.copy, d)
 | 
						|
        self.assertRaises(ValueError, copy.deepcopy, d)
 | 
						|
 | 
						|
    def test_compresspickle(self):
 | 
						|
        for proto in range(pickle.HIGHEST_PROTOCOL + 1):
 | 
						|
            with self.assertRaises((TypeError, pickle.PicklingError)):
 | 
						|
                pickle.dumps(zlib.compressobj(zlib.Z_BEST_COMPRESSION), proto)
 | 
						|
 | 
						|
    def test_decompresspickle(self):
 | 
						|
        for proto in range(pickle.HIGHEST_PROTOCOL + 1):
 | 
						|
            with self.assertRaises((TypeError, pickle.PicklingError)):
 | 
						|
                pickle.dumps(zlib.decompressobj(), proto)
 | 
						|
 | 
						|
    # Memory use of the following functions takes into account overallocation
 | 
						|
 | 
						|
    @bigmemtest(size=_1G + 1024 * 1024, memuse=3)
 | 
						|
    def test_big_compress_buffer(self, size):
 | 
						|
        c = zlib.compressobj(1)
 | 
						|
        compress = lambda s: c.compress(s) + c.flush()
 | 
						|
        self.check_big_compress_buffer(size, compress)
 | 
						|
 | 
						|
    @bigmemtest(size=_1G + 1024 * 1024, memuse=2)
 | 
						|
    def test_big_decompress_buffer(self, size):
 | 
						|
        d = zlib.decompressobj()
 | 
						|
        decompress = lambda s: d.decompress(s) + d.flush()
 | 
						|
        self.check_big_decompress_buffer(size, decompress)
 | 
						|
 | 
						|
    @unittest.skipUnless(sys.maxsize > 2**32, 'requires 64bit platform')
 | 
						|
    @bigmemtest(size=_4G + 100, memuse=4)
 | 
						|
    def test_64bit_compress(self, size):
 | 
						|
        data = b'x' * size
 | 
						|
        co = zlib.compressobj(0)
 | 
						|
        do = zlib.decompressobj()
 | 
						|
        try:
 | 
						|
            comp = co.compress(data) + co.flush()
 | 
						|
            uncomp = do.decompress(comp) + do.flush()
 | 
						|
            self.assertEqual(uncomp, data)
 | 
						|
        finally:
 | 
						|
            comp = uncomp = data = None
 | 
						|
 | 
						|
    @unittest.skipUnless(sys.maxsize > 2**32, 'requires 64bit platform')
 | 
						|
    @bigmemtest(size=_4G + 100, memuse=3)
 | 
						|
    def test_large_unused_data(self, size):
 | 
						|
        data = b'abcdefghijklmnop'
 | 
						|
        unused = b'x' * size
 | 
						|
        comp = zlib.compress(data) + unused
 | 
						|
        do = zlib.decompressobj()
 | 
						|
        try:
 | 
						|
            uncomp = do.decompress(comp) + do.flush()
 | 
						|
            self.assertEqual(unused, do.unused_data)
 | 
						|
            self.assertEqual(uncomp, data)
 | 
						|
        finally:
 | 
						|
            unused = comp = do = None
 | 
						|
 | 
						|
    @unittest.skipUnless(sys.maxsize > 2**32, 'requires 64bit platform')
 | 
						|
    @bigmemtest(size=_4G + 100, memuse=5)
 | 
						|
    def test_large_unconsumed_tail(self, size):
 | 
						|
        data = b'x' * size
 | 
						|
        do = zlib.decompressobj()
 | 
						|
        try:
 | 
						|
            comp = zlib.compress(data, 0)
 | 
						|
            uncomp = do.decompress(comp, 1) + do.flush()
 | 
						|
            self.assertEqual(uncomp, data)
 | 
						|
            self.assertEqual(do.unconsumed_tail, b'')
 | 
						|
        finally:
 | 
						|
            comp = uncomp = data = None
 | 
						|
 | 
						|
    def test_wbits(self):
 | 
						|
        # wbits=0 only supported since zlib v1.2.3.5
 | 
						|
        # Register "1.2.3" as "1.2.3.0"
 | 
						|
        # or "1.2.0-linux","1.2.0.f","1.2.0.f-linux"
 | 
						|
        v = zlib.ZLIB_RUNTIME_VERSION.split('-', 1)[0].split('.')
 | 
						|
        if len(v) < 4:
 | 
						|
            v.append('0')
 | 
						|
        elif not v[-1].isnumeric():
 | 
						|
            v[-1] = '0'
 | 
						|
 | 
						|
        v = tuple(map(int, v))
 | 
						|
        supports_wbits_0 = v >= (1, 2, 3, 5)
 | 
						|
 | 
						|
        co = zlib.compressobj(level=1, wbits=15)
 | 
						|
        zlib15 = co.compress(HAMLET_SCENE) + co.flush()
 | 
						|
        self.assertEqual(zlib.decompress(zlib15, 15), HAMLET_SCENE)
 | 
						|
        if supports_wbits_0:
 | 
						|
            self.assertEqual(zlib.decompress(zlib15, 0), HAMLET_SCENE)
 | 
						|
        self.assertEqual(zlib.decompress(zlib15, 32 + 15), HAMLET_SCENE)
 | 
						|
        with self.assertRaisesRegex(zlib.error, 'invalid window size'):
 | 
						|
            zlib.decompress(zlib15, 14)
 | 
						|
        dco = zlib.decompressobj(wbits=32 + 15)
 | 
						|
        self.assertEqual(dco.decompress(zlib15), HAMLET_SCENE)
 | 
						|
        dco = zlib.decompressobj(wbits=14)
 | 
						|
        with self.assertRaisesRegex(zlib.error, 'invalid window size'):
 | 
						|
            dco.decompress(zlib15)
 | 
						|
 | 
						|
        co = zlib.compressobj(level=1, wbits=9)
 | 
						|
        zlib9 = co.compress(HAMLET_SCENE) + co.flush()
 | 
						|
        self.assertEqual(zlib.decompress(zlib9, 9), HAMLET_SCENE)
 | 
						|
        self.assertEqual(zlib.decompress(zlib9, 15), HAMLET_SCENE)
 | 
						|
        if supports_wbits_0:
 | 
						|
            self.assertEqual(zlib.decompress(zlib9, 0), HAMLET_SCENE)
 | 
						|
        self.assertEqual(zlib.decompress(zlib9, 32 + 9), HAMLET_SCENE)
 | 
						|
        dco = zlib.decompressobj(wbits=32 + 9)
 | 
						|
        self.assertEqual(dco.decompress(zlib9), HAMLET_SCENE)
 | 
						|
 | 
						|
        co = zlib.compressobj(level=1, wbits=-15)
 | 
						|
        deflate15 = co.compress(HAMLET_SCENE) + co.flush()
 | 
						|
        self.assertEqual(zlib.decompress(deflate15, -15), HAMLET_SCENE)
 | 
						|
        dco = zlib.decompressobj(wbits=-15)
 | 
						|
        self.assertEqual(dco.decompress(deflate15), HAMLET_SCENE)
 | 
						|
 | 
						|
        co = zlib.compressobj(level=1, wbits=-9)
 | 
						|
        deflate9 = co.compress(HAMLET_SCENE) + co.flush()
 | 
						|
        self.assertEqual(zlib.decompress(deflate9, -9), HAMLET_SCENE)
 | 
						|
        self.assertEqual(zlib.decompress(deflate9, -15), HAMLET_SCENE)
 | 
						|
        dco = zlib.decompressobj(wbits=-9)
 | 
						|
        self.assertEqual(dco.decompress(deflate9), HAMLET_SCENE)
 | 
						|
 | 
						|
        co = zlib.compressobj(level=1, wbits=16 + 15)
 | 
						|
        gzip = co.compress(HAMLET_SCENE) + co.flush()
 | 
						|
        self.assertEqual(zlib.decompress(gzip, 16 + 15), HAMLET_SCENE)
 | 
						|
        self.assertEqual(zlib.decompress(gzip, 32 + 15), HAMLET_SCENE)
 | 
						|
        dco = zlib.decompressobj(32 + 15)
 | 
						|
        self.assertEqual(dco.decompress(gzip), HAMLET_SCENE)
 | 
						|
 | 
						|
        for wbits in (-15, 15, 31):
 | 
						|
            with self.subTest(wbits=wbits):
 | 
						|
                expected = HAMLET_SCENE
 | 
						|
                actual = zlib.decompress(
 | 
						|
                    zlib.compress(HAMLET_SCENE, wbits=wbits), wbits=wbits
 | 
						|
                )
 | 
						|
                self.assertEqual(expected, actual)
 | 
						|
 | 
						|
def choose_lines(source, number, seed=None, generator=random):
 | 
						|
    """Return a list of number lines randomly chosen from the source"""
 | 
						|
    if seed is not None:
 | 
						|
        generator.seed(seed)
 | 
						|
    sources = source.split('\n')
 | 
						|
    return [generator.choice(sources) for n in range(number)]
 | 
						|
 | 
						|
 | 
						|
HAMLET_SCENE = b"""
 | 
						|
LAERTES
 | 
						|
 | 
						|
       O, fear me not.
 | 
						|
       I stay too long: but here my father comes.
 | 
						|
 | 
						|
       Enter POLONIUS
 | 
						|
 | 
						|
       A double blessing is a double grace,
 | 
						|
       Occasion smiles upon a second leave.
 | 
						|
 | 
						|
LORD POLONIUS
 | 
						|
 | 
						|
       Yet here, Laertes! aboard, aboard, for shame!
 | 
						|
       The wind sits in the shoulder of your sail,
 | 
						|
       And you are stay'd for. There; my blessing with thee!
 | 
						|
       And these few precepts in thy memory
 | 
						|
       See thou character. Give thy thoughts no tongue,
 | 
						|
       Nor any unproportioned thought his act.
 | 
						|
       Be thou familiar, but by no means vulgar.
 | 
						|
       Those friends thou hast, and their adoption tried,
 | 
						|
       Grapple them to thy soul with hoops of steel;
 | 
						|
       But do not dull thy palm with entertainment
 | 
						|
       Of each new-hatch'd, unfledged comrade. Beware
 | 
						|
       Of entrance to a quarrel, but being in,
 | 
						|
       Bear't that the opposed may beware of thee.
 | 
						|
       Give every man thy ear, but few thy voice;
 | 
						|
       Take each man's censure, but reserve thy judgment.
 | 
						|
       Costly thy habit as thy purse can buy,
 | 
						|
       But not express'd in fancy; rich, not gaudy;
 | 
						|
       For the apparel oft proclaims the man,
 | 
						|
       And they in France of the best rank and station
 | 
						|
       Are of a most select and generous chief in that.
 | 
						|
       Neither a borrower nor a lender be;
 | 
						|
       For loan oft loses both itself and friend,
 | 
						|
       And borrowing dulls the edge of husbandry.
 | 
						|
       This above all: to thine ownself be true,
 | 
						|
       And it must follow, as the night the day,
 | 
						|
       Thou canst not then be false to any man.
 | 
						|
       Farewell: my blessing season this in thee!
 | 
						|
 | 
						|
LAERTES
 | 
						|
 | 
						|
       Most humbly do I take my leave, my lord.
 | 
						|
 | 
						|
LORD POLONIUS
 | 
						|
 | 
						|
       The time invites you; go; your servants tend.
 | 
						|
 | 
						|
LAERTES
 | 
						|
 | 
						|
       Farewell, Ophelia; and remember well
 | 
						|
       What I have said to you.
 | 
						|
 | 
						|
OPHELIA
 | 
						|
 | 
						|
       'Tis in my memory lock'd,
 | 
						|
       And you yourself shall keep the key of it.
 | 
						|
 | 
						|
LAERTES
 | 
						|
 | 
						|
       Farewell.
 | 
						|
"""
 | 
						|
 | 
						|
 | 
						|
class CustomInt:
 | 
						|
    def __index__(self):
 | 
						|
        return 100
 | 
						|
 | 
						|
 | 
						|
if __name__ == "__main__":
 | 
						|
    unittest.main()
 |