mirror of
				https://github.com/python/cpython.git
				synced 2025-10-26 11:14:33 +00:00 
			
		
		
		
	 6a44f6eef3
			
		
	
	
		6a44f6eef3
		
			
		
	
	
	
	
		
			
			Deprecate using the __int__() method in implicit conversions of Python numbers to C integers.
		
			
				
	
	
		
			922 lines
		
	
	
	
		
			34 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			922 lines
		
	
	
	
		
			34 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| import unittest
 | |
| from test import support
 | |
| import binascii
 | |
| import copy
 | |
| import pickle
 | |
| import random
 | |
| import sys
 | |
| from test.support import bigmemtest, _1G, _4G
 | |
| 
 | |
| zlib = support.import_module('zlib')
 | |
| 
 | |
| requires_Compress_copy = unittest.skipUnless(
 | |
|         hasattr(zlib.compressobj(), "copy"),
 | |
|         'requires Compress.copy()')
 | |
| requires_Decompress_copy = unittest.skipUnless(
 | |
|         hasattr(zlib.decompressobj(), "copy"),
 | |
|         'requires Decompress.copy()')
 | |
| 
 | |
| 
 | |
| class VersionTestCase(unittest.TestCase):
 | |
| 
 | |
|     def test_library_version(self):
 | |
|         # Test that the major version of the actual library in use matches the
 | |
|         # major version that we were compiled against. We can't guarantee that
 | |
|         # the minor versions will match (even on the machine on which the module
 | |
|         # was compiled), and the API is stable between minor versions, so
 | |
|         # testing only the major versions avoids spurious failures.
 | |
|         self.assertEqual(zlib.ZLIB_RUNTIME_VERSION[0], zlib.ZLIB_VERSION[0])
 | |
| 
 | |
| 
 | |
| class ChecksumTestCase(unittest.TestCase):
 | |
|     # checksum test cases
 | |
|     def test_crc32start(self):
 | |
|         self.assertEqual(zlib.crc32(b""), zlib.crc32(b"", 0))
 | |
|         self.assertTrue(zlib.crc32(b"abc", 0xffffffff))
 | |
| 
 | |
|     def test_crc32empty(self):
 | |
|         self.assertEqual(zlib.crc32(b"", 0), 0)
 | |
|         self.assertEqual(zlib.crc32(b"", 1), 1)
 | |
|         self.assertEqual(zlib.crc32(b"", 432), 432)
 | |
| 
 | |
|     def test_adler32start(self):
 | |
|         self.assertEqual(zlib.adler32(b""), zlib.adler32(b"", 1))
 | |
|         self.assertTrue(zlib.adler32(b"abc", 0xffffffff))
 | |
| 
 | |
|     def test_adler32empty(self):
 | |
|         self.assertEqual(zlib.adler32(b"", 0), 0)
 | |
|         self.assertEqual(zlib.adler32(b"", 1), 1)
 | |
|         self.assertEqual(zlib.adler32(b"", 432), 432)
 | |
| 
 | |
|     def test_penguins(self):
 | |
|         self.assertEqual(zlib.crc32(b"penguin", 0), 0x0e5c1a120)
 | |
|         self.assertEqual(zlib.crc32(b"penguin", 1), 0x43b6aa94)
 | |
|         self.assertEqual(zlib.adler32(b"penguin", 0), 0x0bcf02f6)
 | |
|         self.assertEqual(zlib.adler32(b"penguin", 1), 0x0bd602f7)
 | |
| 
 | |
|         self.assertEqual(zlib.crc32(b"penguin"), zlib.crc32(b"penguin", 0))
 | |
|         self.assertEqual(zlib.adler32(b"penguin"),zlib.adler32(b"penguin",1))
 | |
| 
 | |
|     def test_crc32_adler32_unsigned(self):
 | |
|         foo = b'abcdefghijklmnop'
 | |
|         # explicitly test signed behavior
 | |
|         self.assertEqual(zlib.crc32(foo), 2486878355)
 | |
|         self.assertEqual(zlib.crc32(b'spam'), 1138425661)
 | |
|         self.assertEqual(zlib.adler32(foo+foo), 3573550353)
 | |
|         self.assertEqual(zlib.adler32(b'spam'), 72286642)
 | |
| 
 | |
|     def test_same_as_binascii_crc32(self):
 | |
|         foo = b'abcdefghijklmnop'
 | |
|         crc = 2486878355
 | |
|         self.assertEqual(binascii.crc32(foo), crc)
 | |
|         self.assertEqual(zlib.crc32(foo), crc)
 | |
|         self.assertEqual(binascii.crc32(b'spam'), zlib.crc32(b'spam'))
 | |
| 
 | |
| 
 | |
| # Issue #10276 - check that inputs >=4 GiB are handled correctly.
 | |
| class ChecksumBigBufferTestCase(unittest.TestCase):
 | |
| 
 | |
|     @bigmemtest(size=_4G + 4, memuse=1, dry_run=False)
 | |
|     def test_big_buffer(self, size):
 | |
|         data = b"nyan" * (_1G + 1)
 | |
|         self.assertEqual(zlib.crc32(data), 1044521549)
 | |
|         self.assertEqual(zlib.adler32(data), 2256789997)
 | |
| 
 | |
| 
 | |
| class ExceptionTestCase(unittest.TestCase):
 | |
|     # make sure we generate some expected errors
 | |
|     def test_badlevel(self):
 | |
|         # specifying compression level out of range causes an error
 | |
|         # (but -1 is Z_DEFAULT_COMPRESSION and apparently the zlib
 | |
|         # accepts 0 too)
 | |
|         self.assertRaises(zlib.error, zlib.compress, b'ERROR', 10)
 | |
| 
 | |
|     def test_badargs(self):
 | |
|         self.assertRaises(TypeError, zlib.adler32)
 | |
|         self.assertRaises(TypeError, zlib.crc32)
 | |
|         self.assertRaises(TypeError, zlib.compress)
 | |
|         self.assertRaises(TypeError, zlib.decompress)
 | |
|         for arg in (42, None, '', 'abc', (), []):
 | |
|             self.assertRaises(TypeError, zlib.adler32, arg)
 | |
|             self.assertRaises(TypeError, zlib.crc32, arg)
 | |
|             self.assertRaises(TypeError, zlib.compress, arg)
 | |
|             self.assertRaises(TypeError, zlib.decompress, arg)
 | |
| 
 | |
|     def test_badcompressobj(self):
 | |
|         # verify failure on building compress object with bad params
 | |
|         self.assertRaises(ValueError, zlib.compressobj, 1, zlib.DEFLATED, 0)
 | |
|         # specifying total bits too large causes an error
 | |
|         self.assertRaises(ValueError,
 | |
|                 zlib.compressobj, 1, zlib.DEFLATED, zlib.MAX_WBITS + 1)
 | |
| 
 | |
|     def test_baddecompressobj(self):
 | |
|         # verify failure on building decompress object with bad params
 | |
|         self.assertRaises(ValueError, zlib.decompressobj, -1)
 | |
| 
 | |
|     def test_decompressobj_badflush(self):
 | |
|         # verify failure on calling decompressobj.flush with bad params
 | |
|         self.assertRaises(ValueError, zlib.decompressobj().flush, 0)
 | |
|         self.assertRaises(ValueError, zlib.decompressobj().flush, -1)
 | |
| 
 | |
|     @support.cpython_only
 | |
|     def test_overflow(self):
 | |
|         with self.assertRaisesRegex(OverflowError, 'int too large'):
 | |
|             zlib.decompress(b'', 15, sys.maxsize + 1)
 | |
|         with self.assertRaisesRegex(OverflowError, 'int too large'):
 | |
|             zlib.decompressobj().decompress(b'', sys.maxsize + 1)
 | |
|         with self.assertRaisesRegex(OverflowError, 'int too large'):
 | |
|             zlib.decompressobj().flush(sys.maxsize + 1)
 | |
| 
 | |
| 
 | |
| class BaseCompressTestCase(object):
 | |
|     def check_big_compress_buffer(self, size, compress_func):
 | |
|         _1M = 1024 * 1024
 | |
|         # Generate 10 MiB worth of random, and expand it by repeating it.
 | |
|         # The assumption is that zlib's memory is not big enough to exploit
 | |
|         # such spread out redundancy.
 | |
|         data = b''.join([random.getrandbits(8 * _1M).to_bytes(_1M, 'little')
 | |
|                         for i in range(10)])
 | |
|         data = data * (size // len(data) + 1)
 | |
|         try:
 | |
|             compress_func(data)
 | |
|         finally:
 | |
|             # Release memory
 | |
|             data = None
 | |
| 
 | |
|     def check_big_decompress_buffer(self, size, decompress_func):
 | |
|         data = b'x' * size
 | |
|         try:
 | |
|             compressed = zlib.compress(data, 1)
 | |
|         finally:
 | |
|             # Release memory
 | |
|             data = None
 | |
|         data = decompress_func(compressed)
 | |
|         # Sanity check
 | |
|         try:
 | |
|             self.assertEqual(len(data), size)
 | |
|             self.assertEqual(len(data.strip(b'x')), 0)
 | |
|         finally:
 | |
|             data = None
 | |
| 
 | |
| 
 | |
| class CompressTestCase(BaseCompressTestCase, unittest.TestCase):
 | |
|     # Test compression in one go (whole message compression)
 | |
|     def test_speech(self):
 | |
|         x = zlib.compress(HAMLET_SCENE)
 | |
|         self.assertEqual(zlib.decompress(x), HAMLET_SCENE)
 | |
| 
 | |
|     def test_keywords(self):
 | |
|         x = zlib.compress(HAMLET_SCENE, level=3)
 | |
|         self.assertEqual(zlib.decompress(x), HAMLET_SCENE)
 | |
|         with self.assertRaises(TypeError):
 | |
|             zlib.compress(data=HAMLET_SCENE, level=3)
 | |
|         self.assertEqual(zlib.decompress(x,
 | |
|                                          wbits=zlib.MAX_WBITS,
 | |
|                                          bufsize=zlib.DEF_BUF_SIZE),
 | |
|                          HAMLET_SCENE)
 | |
| 
 | |
|     def test_speech128(self):
 | |
|         # compress more data
 | |
|         data = HAMLET_SCENE * 128
 | |
|         x = zlib.compress(data)
 | |
|         self.assertEqual(zlib.compress(bytearray(data)), x)
 | |
|         for ob in x, bytearray(x):
 | |
|             self.assertEqual(zlib.decompress(ob), data)
 | |
| 
 | |
|     def test_incomplete_stream(self):
 | |
|         # A useful error message is given
 | |
|         x = zlib.compress(HAMLET_SCENE)
 | |
|         self.assertRaisesRegex(zlib.error,
 | |
|             "Error -5 while decompressing data: incomplete or truncated stream",
 | |
|             zlib.decompress, x[:-1])
 | |
| 
 | |
|     # Memory use of the following functions takes into account overallocation
 | |
| 
 | |
|     @bigmemtest(size=_1G + 1024 * 1024, memuse=3)
 | |
|     def test_big_compress_buffer(self, size):
 | |
|         compress = lambda s: zlib.compress(s, 1)
 | |
|         self.check_big_compress_buffer(size, compress)
 | |
| 
 | |
|     @bigmemtest(size=_1G + 1024 * 1024, memuse=2)
 | |
|     def test_big_decompress_buffer(self, size):
 | |
|         self.check_big_decompress_buffer(size, zlib.decompress)
 | |
| 
 | |
|     @bigmemtest(size=_4G, memuse=1)
 | |
|     def test_large_bufsize(self, size):
 | |
|         # Test decompress(bufsize) parameter greater than the internal limit
 | |
|         data = HAMLET_SCENE * 10
 | |
|         compressed = zlib.compress(data, 1)
 | |
|         self.assertEqual(zlib.decompress(compressed, 15, size), data)
 | |
| 
 | |
|     def test_custom_bufsize(self):
 | |
|         data = HAMLET_SCENE * 10
 | |
|         compressed = zlib.compress(data, 1)
 | |
|         self.assertEqual(zlib.decompress(compressed, 15, CustomInt()), data)
 | |
| 
 | |
|     @unittest.skipUnless(sys.maxsize > 2**32, 'requires 64bit platform')
 | |
|     @bigmemtest(size=_4G + 100, memuse=4)
 | |
|     def test_64bit_compress(self, size):
 | |
|         data = b'x' * size
 | |
|         try:
 | |
|             comp = zlib.compress(data, 0)
 | |
|             self.assertEqual(zlib.decompress(comp), data)
 | |
|         finally:
 | |
|             comp = data = None
 | |
| 
 | |
| 
 | |
| class CompressObjectTestCase(BaseCompressTestCase, unittest.TestCase):
 | |
|     # Test compression object
 | |
|     def test_pair(self):
 | |
|         # straightforward compress/decompress objects
 | |
|         datasrc = HAMLET_SCENE * 128
 | |
|         datazip = zlib.compress(datasrc)
 | |
|         # should compress both bytes and bytearray data
 | |
|         for data in (datasrc, bytearray(datasrc)):
 | |
|             co = zlib.compressobj()
 | |
|             x1 = co.compress(data)
 | |
|             x2 = co.flush()
 | |
|             self.assertRaises(zlib.error, co.flush) # second flush should not work
 | |
|             self.assertEqual(x1 + x2, datazip)
 | |
|         for v1, v2 in ((x1, x2), (bytearray(x1), bytearray(x2))):
 | |
|             dco = zlib.decompressobj()
 | |
|             y1 = dco.decompress(v1 + v2)
 | |
|             y2 = dco.flush()
 | |
|             self.assertEqual(data, y1 + y2)
 | |
|             self.assertIsInstance(dco.unconsumed_tail, bytes)
 | |
|             self.assertIsInstance(dco.unused_data, bytes)
 | |
| 
 | |
|     def test_keywords(self):
 | |
|         level = 2
 | |
|         method = zlib.DEFLATED
 | |
|         wbits = -12
 | |
|         memLevel = 9
 | |
|         strategy = zlib.Z_FILTERED
 | |
|         co = zlib.compressobj(level=level,
 | |
|                               method=method,
 | |
|                               wbits=wbits,
 | |
|                               memLevel=memLevel,
 | |
|                               strategy=strategy,
 | |
|                               zdict=b"")
 | |
|         do = zlib.decompressobj(wbits=wbits, zdict=b"")
 | |
|         with self.assertRaises(TypeError):
 | |
|             co.compress(data=HAMLET_SCENE)
 | |
|         with self.assertRaises(TypeError):
 | |
|             do.decompress(data=zlib.compress(HAMLET_SCENE))
 | |
|         x = co.compress(HAMLET_SCENE) + co.flush()
 | |
|         y = do.decompress(x, max_length=len(HAMLET_SCENE)) + do.flush()
 | |
|         self.assertEqual(HAMLET_SCENE, y)
 | |
| 
 | |
|     def test_compressoptions(self):
 | |
|         # specify lots of options to compressobj()
 | |
|         level = 2
 | |
|         method = zlib.DEFLATED
 | |
|         wbits = -12
 | |
|         memLevel = 9
 | |
|         strategy = zlib.Z_FILTERED
 | |
|         co = zlib.compressobj(level, method, wbits, memLevel, strategy)
 | |
|         x1 = co.compress(HAMLET_SCENE)
 | |
|         x2 = co.flush()
 | |
|         dco = zlib.decompressobj(wbits)
 | |
|         y1 = dco.decompress(x1 + x2)
 | |
|         y2 = dco.flush()
 | |
|         self.assertEqual(HAMLET_SCENE, y1 + y2)
 | |
| 
 | |
|     def test_compressincremental(self):
 | |
|         # compress object in steps, decompress object as one-shot
 | |
|         data = HAMLET_SCENE * 128
 | |
|         co = zlib.compressobj()
 | |
|         bufs = []
 | |
|         for i in range(0, len(data), 256):
 | |
|             bufs.append(co.compress(data[i:i+256]))
 | |
|         bufs.append(co.flush())
 | |
|         combuf = b''.join(bufs)
 | |
| 
 | |
|         dco = zlib.decompressobj()
 | |
|         y1 = dco.decompress(b''.join(bufs))
 | |
|         y2 = dco.flush()
 | |
|         self.assertEqual(data, y1 + y2)
 | |
| 
 | |
|     def test_decompinc(self, flush=False, source=None, cx=256, dcx=64):
 | |
|         # compress object in steps, decompress object in steps
 | |
|         source = source or HAMLET_SCENE
 | |
|         data = source * 128
 | |
|         co = zlib.compressobj()
 | |
|         bufs = []
 | |
|         for i in range(0, len(data), cx):
 | |
|             bufs.append(co.compress(data[i:i+cx]))
 | |
|         bufs.append(co.flush())
 | |
|         combuf = b''.join(bufs)
 | |
| 
 | |
|         decombuf = zlib.decompress(combuf)
 | |
|         # Test type of return value
 | |
|         self.assertIsInstance(decombuf, bytes)
 | |
| 
 | |
|         self.assertEqual(data, decombuf)
 | |
| 
 | |
|         dco = zlib.decompressobj()
 | |
|         bufs = []
 | |
|         for i in range(0, len(combuf), dcx):
 | |
|             bufs.append(dco.decompress(combuf[i:i+dcx]))
 | |
|             self.assertEqual(b'', dco.unconsumed_tail, ########
 | |
|                              "(A) uct should be b'': not %d long" %
 | |
|                                        len(dco.unconsumed_tail))
 | |
|             self.assertEqual(b'', dco.unused_data)
 | |
|         if flush:
 | |
|             bufs.append(dco.flush())
 | |
|         else:
 | |
|             while True:
 | |
|                 chunk = dco.decompress(b'')
 | |
|                 if chunk:
 | |
|                     bufs.append(chunk)
 | |
|                 else:
 | |
|                     break
 | |
|         self.assertEqual(b'', dco.unconsumed_tail, ########
 | |
|                          "(B) uct should be b'': not %d long" %
 | |
|                                        len(dco.unconsumed_tail))
 | |
|         self.assertEqual(b'', dco.unused_data)
 | |
|         self.assertEqual(data, b''.join(bufs))
 | |
|         # Failure means: "decompressobj with init options failed"
 | |
| 
 | |
|     def test_decompincflush(self):
 | |
|         self.test_decompinc(flush=True)
 | |
| 
 | |
|     def test_decompimax(self, source=None, cx=256, dcx=64):
 | |
|         # compress in steps, decompress in length-restricted steps
 | |
|         source = source or HAMLET_SCENE
 | |
|         # Check a decompression object with max_length specified
 | |
|         data = source * 128
 | |
|         co = zlib.compressobj()
 | |
|         bufs = []
 | |
|         for i in range(0, len(data), cx):
 | |
|             bufs.append(co.compress(data[i:i+cx]))
 | |
|         bufs.append(co.flush())
 | |
|         combuf = b''.join(bufs)
 | |
|         self.assertEqual(data, zlib.decompress(combuf),
 | |
|                          'compressed data failure')
 | |
| 
 | |
|         dco = zlib.decompressobj()
 | |
|         bufs = []
 | |
|         cb = combuf
 | |
|         while cb:
 | |
|             #max_length = 1 + len(cb)//10
 | |
|             chunk = dco.decompress(cb, dcx)
 | |
|             self.assertFalse(len(chunk) > dcx,
 | |
|                     'chunk too big (%d>%d)' % (len(chunk), dcx))
 | |
|             bufs.append(chunk)
 | |
|             cb = dco.unconsumed_tail
 | |
|         bufs.append(dco.flush())
 | |
|         self.assertEqual(data, b''.join(bufs), 'Wrong data retrieved')
 | |
| 
 | |
|     def test_decompressmaxlen(self, flush=False):
 | |
|         # Check a decompression object with max_length specified
 | |
|         data = HAMLET_SCENE * 128
 | |
|         co = zlib.compressobj()
 | |
|         bufs = []
 | |
|         for i in range(0, len(data), 256):
 | |
|             bufs.append(co.compress(data[i:i+256]))
 | |
|         bufs.append(co.flush())
 | |
|         combuf = b''.join(bufs)
 | |
|         self.assertEqual(data, zlib.decompress(combuf),
 | |
|                          'compressed data failure')
 | |
| 
 | |
|         dco = zlib.decompressobj()
 | |
|         bufs = []
 | |
|         cb = combuf
 | |
|         while cb:
 | |
|             max_length = 1 + len(cb)//10
 | |
|             chunk = dco.decompress(cb, max_length)
 | |
|             self.assertFalse(len(chunk) > max_length,
 | |
|                         'chunk too big (%d>%d)' % (len(chunk),max_length))
 | |
|             bufs.append(chunk)
 | |
|             cb = dco.unconsumed_tail
 | |
|         if flush:
 | |
|             bufs.append(dco.flush())
 | |
|         else:
 | |
|             while chunk:
 | |
|                 chunk = dco.decompress(b'', max_length)
 | |
|                 self.assertFalse(len(chunk) > max_length,
 | |
|                             'chunk too big (%d>%d)' % (len(chunk),max_length))
 | |
|                 bufs.append(chunk)
 | |
|         self.assertEqual(data, b''.join(bufs), 'Wrong data retrieved')
 | |
| 
 | |
|     def test_decompressmaxlenflush(self):
 | |
|         self.test_decompressmaxlen(flush=True)
 | |
| 
 | |
|     def test_maxlenmisc(self):
 | |
|         # Misc tests of max_length
 | |
|         dco = zlib.decompressobj()
 | |
|         self.assertRaises(ValueError, dco.decompress, b"", -1)
 | |
|         self.assertEqual(b'', dco.unconsumed_tail)
 | |
| 
 | |
|     def test_maxlen_large(self):
 | |
|         # Sizes up to sys.maxsize should be accepted, although zlib is
 | |
|         # internally limited to expressing sizes with unsigned int
 | |
|         data = HAMLET_SCENE * 10
 | |
|         self.assertGreater(len(data), zlib.DEF_BUF_SIZE)
 | |
|         compressed = zlib.compress(data, 1)
 | |
|         dco = zlib.decompressobj()
 | |
|         self.assertEqual(dco.decompress(compressed, sys.maxsize), data)
 | |
| 
 | |
|     def test_maxlen_custom(self):
 | |
|         data = HAMLET_SCENE * 10
 | |
|         compressed = zlib.compress(data, 1)
 | |
|         dco = zlib.decompressobj()
 | |
|         self.assertEqual(dco.decompress(compressed, CustomInt()), data[:100])
 | |
| 
 | |
|     def test_clear_unconsumed_tail(self):
 | |
|         # Issue #12050: calling decompress() without providing max_length
 | |
|         # should clear the unconsumed_tail attribute.
 | |
|         cdata = b"x\x9cKLJ\x06\x00\x02M\x01"    # "abc"
 | |
|         dco = zlib.decompressobj()
 | |
|         ddata = dco.decompress(cdata, 1)
 | |
|         ddata += dco.decompress(dco.unconsumed_tail)
 | |
|         self.assertEqual(dco.unconsumed_tail, b"")
 | |
| 
 | |
|     def test_flushes(self):
 | |
|         # Test flush() with the various options, using all the
 | |
|         # different levels in order to provide more variations.
 | |
|         sync_opt = ['Z_NO_FLUSH', 'Z_SYNC_FLUSH', 'Z_FULL_FLUSH',
 | |
|                     'Z_PARTIAL_FLUSH']
 | |
| 
 | |
|         ver = tuple(int(v) for v in zlib.ZLIB_RUNTIME_VERSION.split('.'))
 | |
|         # Z_BLOCK has a known failure prior to 1.2.5.3
 | |
|         if ver >= (1, 2, 5, 3):
 | |
|             sync_opt.append('Z_BLOCK')
 | |
| 
 | |
|         sync_opt = [getattr(zlib, opt) for opt in sync_opt
 | |
|                     if hasattr(zlib, opt)]
 | |
|         data = HAMLET_SCENE * 8
 | |
| 
 | |
|         for sync in sync_opt:
 | |
|             for level in range(10):
 | |
|                 try:
 | |
|                     obj = zlib.compressobj( level )
 | |
|                     a = obj.compress( data[:3000] )
 | |
|                     b = obj.flush( sync )
 | |
|                     c = obj.compress( data[3000:] )
 | |
|                     d = obj.flush()
 | |
|                 except:
 | |
|                     print("Error for flush mode={}, level={}"
 | |
|                           .format(sync, level))
 | |
|                     raise
 | |
|                 self.assertEqual(zlib.decompress(b''.join([a,b,c,d])),
 | |
|                                  data, ("Decompress failed: flush "
 | |
|                                         "mode=%i, level=%i") % (sync, level))
 | |
|                 del obj
 | |
| 
 | |
|     @unittest.skipUnless(hasattr(zlib, 'Z_SYNC_FLUSH'),
 | |
|                          'requires zlib.Z_SYNC_FLUSH')
 | |
|     def test_odd_flush(self):
 | |
|         # Test for odd flushing bugs noted in 2.0, and hopefully fixed in 2.1
 | |
|         import random
 | |
|         # Testing on 17K of "random" data
 | |
| 
 | |
|         # Create compressor and decompressor objects
 | |
|         co = zlib.compressobj(zlib.Z_BEST_COMPRESSION)
 | |
|         dco = zlib.decompressobj()
 | |
| 
 | |
|         # Try 17K of data
 | |
|         # generate random data stream
 | |
|         try:
 | |
|             # In 2.3 and later, WichmannHill is the RNG of the bug report
 | |
|             gen = random.WichmannHill()
 | |
|         except AttributeError:
 | |
|             try:
 | |
|                 # 2.2 called it Random
 | |
|                 gen = random.Random()
 | |
|             except AttributeError:
 | |
|                 # others might simply have a single RNG
 | |
|                 gen = random
 | |
|         gen.seed(1)
 | |
|         data = genblock(1, 17 * 1024, generator=gen)
 | |
| 
 | |
|         # compress, sync-flush, and decompress
 | |
|         first = co.compress(data)
 | |
|         second = co.flush(zlib.Z_SYNC_FLUSH)
 | |
|         expanded = dco.decompress(first + second)
 | |
| 
 | |
|         # if decompressed data is different from the input data, choke.
 | |
|         self.assertEqual(expanded, data, "17K random source doesn't match")
 | |
| 
 | |
|     def test_empty_flush(self):
 | |
|         # Test that calling .flush() on unused objects works.
 | |
|         # (Bug #1083110 -- calling .flush() on decompress objects
 | |
|         # caused a core dump.)
 | |
| 
 | |
|         co = zlib.compressobj(zlib.Z_BEST_COMPRESSION)
 | |
|         self.assertTrue(co.flush())  # Returns a zlib header
 | |
|         dco = zlib.decompressobj()
 | |
|         self.assertEqual(dco.flush(), b"") # Returns nothing
 | |
| 
 | |
|     def test_dictionary(self):
 | |
|         h = HAMLET_SCENE
 | |
|         # Build a simulated dictionary out of the words in HAMLET.
 | |
|         words = h.split()
 | |
|         random.shuffle(words)
 | |
|         zdict = b''.join(words)
 | |
|         # Use it to compress HAMLET.
 | |
|         co = zlib.compressobj(zdict=zdict)
 | |
|         cd = co.compress(h) + co.flush()
 | |
|         # Verify that it will decompress with the dictionary.
 | |
|         dco = zlib.decompressobj(zdict=zdict)
 | |
|         self.assertEqual(dco.decompress(cd) + dco.flush(), h)
 | |
|         # Verify that it fails when not given the dictionary.
 | |
|         dco = zlib.decompressobj()
 | |
|         self.assertRaises(zlib.error, dco.decompress, cd)
 | |
| 
 | |
|     def test_dictionary_streaming(self):
 | |
|         # This simulates the reuse of a compressor object for compressing
 | |
|         # several separate data streams.
 | |
|         co = zlib.compressobj(zdict=HAMLET_SCENE)
 | |
|         do = zlib.decompressobj(zdict=HAMLET_SCENE)
 | |
|         piece = HAMLET_SCENE[1000:1500]
 | |
|         d0 = co.compress(piece) + co.flush(zlib.Z_SYNC_FLUSH)
 | |
|         d1 = co.compress(piece[100:]) + co.flush(zlib.Z_SYNC_FLUSH)
 | |
|         d2 = co.compress(piece[:-100]) + co.flush(zlib.Z_SYNC_FLUSH)
 | |
|         self.assertEqual(do.decompress(d0), piece)
 | |
|         self.assertEqual(do.decompress(d1), piece[100:])
 | |
|         self.assertEqual(do.decompress(d2), piece[:-100])
 | |
| 
 | |
|     def test_decompress_incomplete_stream(self):
 | |
|         # This is 'foo', deflated
 | |
|         x = b'x\x9cK\xcb\xcf\x07\x00\x02\x82\x01E'
 | |
|         # For the record
 | |
|         self.assertEqual(zlib.decompress(x), b'foo')
 | |
|         self.assertRaises(zlib.error, zlib.decompress, x[:-5])
 | |
|         # Omitting the stream end works with decompressor objects
 | |
|         # (see issue #8672).
 | |
|         dco = zlib.decompressobj()
 | |
|         y = dco.decompress(x[:-5])
 | |
|         y += dco.flush()
 | |
|         self.assertEqual(y, b'foo')
 | |
| 
 | |
|     def test_decompress_eof(self):
 | |
|         x = b'x\x9cK\xcb\xcf\x07\x00\x02\x82\x01E'  # 'foo'
 | |
|         dco = zlib.decompressobj()
 | |
|         self.assertFalse(dco.eof)
 | |
|         dco.decompress(x[:-5])
 | |
|         self.assertFalse(dco.eof)
 | |
|         dco.decompress(x[-5:])
 | |
|         self.assertTrue(dco.eof)
 | |
|         dco.flush()
 | |
|         self.assertTrue(dco.eof)
 | |
| 
 | |
|     def test_decompress_eof_incomplete_stream(self):
 | |
|         x = b'x\x9cK\xcb\xcf\x07\x00\x02\x82\x01E'  # 'foo'
 | |
|         dco = zlib.decompressobj()
 | |
|         self.assertFalse(dco.eof)
 | |
|         dco.decompress(x[:-5])
 | |
|         self.assertFalse(dco.eof)
 | |
|         dco.flush()
 | |
|         self.assertFalse(dco.eof)
 | |
| 
 | |
|     def test_decompress_unused_data(self):
 | |
|         # Repeated calls to decompress() after EOF should accumulate data in
 | |
|         # dco.unused_data, instead of just storing the arg to the last call.
 | |
|         source = b'abcdefghijklmnopqrstuvwxyz'
 | |
|         remainder = b'0123456789'
 | |
|         y = zlib.compress(source)
 | |
|         x = y + remainder
 | |
|         for maxlen in 0, 1000:
 | |
|             for step in 1, 2, len(y), len(x):
 | |
|                 dco = zlib.decompressobj()
 | |
|                 data = b''
 | |
|                 for i in range(0, len(x), step):
 | |
|                     if i < len(y):
 | |
|                         self.assertEqual(dco.unused_data, b'')
 | |
|                     if maxlen == 0:
 | |
|                         data += dco.decompress(x[i : i + step])
 | |
|                         self.assertEqual(dco.unconsumed_tail, b'')
 | |
|                     else:
 | |
|                         data += dco.decompress(
 | |
|                                 dco.unconsumed_tail + x[i : i + step], maxlen)
 | |
|                 data += dco.flush()
 | |
|                 self.assertTrue(dco.eof)
 | |
|                 self.assertEqual(data, source)
 | |
|                 self.assertEqual(dco.unconsumed_tail, b'')
 | |
|                 self.assertEqual(dco.unused_data, remainder)
 | |
| 
 | |
|     # issue27164
 | |
|     def test_decompress_raw_with_dictionary(self):
 | |
|         zdict = b'abcdefghijklmnopqrstuvwxyz'
 | |
|         co = zlib.compressobj(wbits=-zlib.MAX_WBITS, zdict=zdict)
 | |
|         comp = co.compress(zdict) + co.flush()
 | |
|         dco = zlib.decompressobj(wbits=-zlib.MAX_WBITS, zdict=zdict)
 | |
|         uncomp = dco.decompress(comp) + dco.flush()
 | |
|         self.assertEqual(zdict, uncomp)
 | |
| 
 | |
|     def test_flush_with_freed_input(self):
 | |
|         # Issue #16411: decompressor accesses input to last decompress() call
 | |
|         # in flush(), even if this object has been freed in the meanwhile.
 | |
|         input1 = b'abcdefghijklmnopqrstuvwxyz'
 | |
|         input2 = b'QWERTYUIOPASDFGHJKLZXCVBNM'
 | |
|         data = zlib.compress(input1)
 | |
|         dco = zlib.decompressobj()
 | |
|         dco.decompress(data, 1)
 | |
|         del data
 | |
|         data = zlib.compress(input2)
 | |
|         self.assertEqual(dco.flush(), input1[1:])
 | |
| 
 | |
|     @bigmemtest(size=_4G, memuse=1)
 | |
|     def test_flush_large_length(self, size):
 | |
|         # Test flush(length) parameter greater than internal limit UINT_MAX
 | |
|         input = HAMLET_SCENE * 10
 | |
|         data = zlib.compress(input, 1)
 | |
|         dco = zlib.decompressobj()
 | |
|         dco.decompress(data, 1)
 | |
|         self.assertEqual(dco.flush(size), input[1:])
 | |
| 
 | |
|     def test_flush_custom_length(self):
 | |
|         input = HAMLET_SCENE * 10
 | |
|         data = zlib.compress(input, 1)
 | |
|         dco = zlib.decompressobj()
 | |
|         dco.decompress(data, 1)
 | |
|         self.assertEqual(dco.flush(CustomInt()), input[1:])
 | |
| 
 | |
|     @requires_Compress_copy
 | |
|     def test_compresscopy(self):
 | |
|         # Test copying a compression object
 | |
|         data0 = HAMLET_SCENE
 | |
|         data1 = bytes(str(HAMLET_SCENE, "ascii").swapcase(), "ascii")
 | |
|         for func in lambda c: c.copy(), copy.copy, copy.deepcopy:
 | |
|             c0 = zlib.compressobj(zlib.Z_BEST_COMPRESSION)
 | |
|             bufs0 = []
 | |
|             bufs0.append(c0.compress(data0))
 | |
| 
 | |
|             c1 = func(c0)
 | |
|             bufs1 = bufs0[:]
 | |
| 
 | |
|             bufs0.append(c0.compress(data0))
 | |
|             bufs0.append(c0.flush())
 | |
|             s0 = b''.join(bufs0)
 | |
| 
 | |
|             bufs1.append(c1.compress(data1))
 | |
|             bufs1.append(c1.flush())
 | |
|             s1 = b''.join(bufs1)
 | |
| 
 | |
|             self.assertEqual(zlib.decompress(s0),data0+data0)
 | |
|             self.assertEqual(zlib.decompress(s1),data0+data1)
 | |
| 
 | |
|     @requires_Compress_copy
 | |
|     def test_badcompresscopy(self):
 | |
|         # Test copying a compression object in an inconsistent state
 | |
|         c = zlib.compressobj()
 | |
|         c.compress(HAMLET_SCENE)
 | |
|         c.flush()
 | |
|         self.assertRaises(ValueError, c.copy)
 | |
|         self.assertRaises(ValueError, copy.copy, c)
 | |
|         self.assertRaises(ValueError, copy.deepcopy, c)
 | |
| 
 | |
|     @requires_Decompress_copy
 | |
|     def test_decompresscopy(self):
 | |
|         # Test copying a decompression object
 | |
|         data = HAMLET_SCENE
 | |
|         comp = zlib.compress(data)
 | |
|         # Test type of return value
 | |
|         self.assertIsInstance(comp, bytes)
 | |
| 
 | |
|         for func in lambda c: c.copy(), copy.copy, copy.deepcopy:
 | |
|             d0 = zlib.decompressobj()
 | |
|             bufs0 = []
 | |
|             bufs0.append(d0.decompress(comp[:32]))
 | |
| 
 | |
|             d1 = func(d0)
 | |
|             bufs1 = bufs0[:]
 | |
| 
 | |
|             bufs0.append(d0.decompress(comp[32:]))
 | |
|             s0 = b''.join(bufs0)
 | |
| 
 | |
|             bufs1.append(d1.decompress(comp[32:]))
 | |
|             s1 = b''.join(bufs1)
 | |
| 
 | |
|             self.assertEqual(s0,s1)
 | |
|             self.assertEqual(s0,data)
 | |
| 
 | |
|     @requires_Decompress_copy
 | |
|     def test_baddecompresscopy(self):
 | |
|         # Test copying a compression object in an inconsistent state
 | |
|         data = zlib.compress(HAMLET_SCENE)
 | |
|         d = zlib.decompressobj()
 | |
|         d.decompress(data)
 | |
|         d.flush()
 | |
|         self.assertRaises(ValueError, d.copy)
 | |
|         self.assertRaises(ValueError, copy.copy, d)
 | |
|         self.assertRaises(ValueError, copy.deepcopy, d)
 | |
| 
 | |
|     def test_compresspickle(self):
 | |
|         for proto in range(pickle.HIGHEST_PROTOCOL + 1):
 | |
|             with self.assertRaises((TypeError, pickle.PicklingError)):
 | |
|                 pickle.dumps(zlib.compressobj(zlib.Z_BEST_COMPRESSION), proto)
 | |
| 
 | |
|     def test_decompresspickle(self):
 | |
|         for proto in range(pickle.HIGHEST_PROTOCOL + 1):
 | |
|             with self.assertRaises((TypeError, pickle.PicklingError)):
 | |
|                 pickle.dumps(zlib.decompressobj(), proto)
 | |
| 
 | |
|     # Memory use of the following functions takes into account overallocation
 | |
| 
 | |
|     @bigmemtest(size=_1G + 1024 * 1024, memuse=3)
 | |
|     def test_big_compress_buffer(self, size):
 | |
|         c = zlib.compressobj(1)
 | |
|         compress = lambda s: c.compress(s) + c.flush()
 | |
|         self.check_big_compress_buffer(size, compress)
 | |
| 
 | |
|     @bigmemtest(size=_1G + 1024 * 1024, memuse=2)
 | |
|     def test_big_decompress_buffer(self, size):
 | |
|         d = zlib.decompressobj()
 | |
|         decompress = lambda s: d.decompress(s) + d.flush()
 | |
|         self.check_big_decompress_buffer(size, decompress)
 | |
| 
 | |
|     @unittest.skipUnless(sys.maxsize > 2**32, 'requires 64bit platform')
 | |
|     @bigmemtest(size=_4G + 100, memuse=4)
 | |
|     def test_64bit_compress(self, size):
 | |
|         data = b'x' * size
 | |
|         co = zlib.compressobj(0)
 | |
|         do = zlib.decompressobj()
 | |
|         try:
 | |
|             comp = co.compress(data) + co.flush()
 | |
|             uncomp = do.decompress(comp) + do.flush()
 | |
|             self.assertEqual(uncomp, data)
 | |
|         finally:
 | |
|             comp = uncomp = data = None
 | |
| 
 | |
|     @unittest.skipUnless(sys.maxsize > 2**32, 'requires 64bit platform')
 | |
|     @bigmemtest(size=_4G + 100, memuse=3)
 | |
|     def test_large_unused_data(self, size):
 | |
|         data = b'abcdefghijklmnop'
 | |
|         unused = b'x' * size
 | |
|         comp = zlib.compress(data) + unused
 | |
|         do = zlib.decompressobj()
 | |
|         try:
 | |
|             uncomp = do.decompress(comp) + do.flush()
 | |
|             self.assertEqual(unused, do.unused_data)
 | |
|             self.assertEqual(uncomp, data)
 | |
|         finally:
 | |
|             unused = comp = do = None
 | |
| 
 | |
|     @unittest.skipUnless(sys.maxsize > 2**32, 'requires 64bit platform')
 | |
|     @bigmemtest(size=_4G + 100, memuse=5)
 | |
|     def test_large_unconsumed_tail(self, size):
 | |
|         data = b'x' * size
 | |
|         do = zlib.decompressobj()
 | |
|         try:
 | |
|             comp = zlib.compress(data, 0)
 | |
|             uncomp = do.decompress(comp, 1) + do.flush()
 | |
|             self.assertEqual(uncomp, data)
 | |
|             self.assertEqual(do.unconsumed_tail, b'')
 | |
|         finally:
 | |
|             comp = uncomp = data = None
 | |
| 
 | |
|     def test_wbits(self):
 | |
|         # wbits=0 only supported since zlib v1.2.3.5
 | |
|         # Register "1.2.3" as "1.2.3.0"
 | |
|         # or "1.2.0-linux","1.2.0.f","1.2.0.f-linux"
 | |
|         v = zlib.ZLIB_RUNTIME_VERSION.split('-', 1)[0].split('.')
 | |
|         if len(v) < 4:
 | |
|             v.append('0')
 | |
|         elif not v[-1].isnumeric():
 | |
|             v[-1] = '0'
 | |
| 
 | |
|         v = tuple(map(int, v))
 | |
|         supports_wbits_0 = v >= (1, 2, 3, 5)
 | |
| 
 | |
|         co = zlib.compressobj(level=1, wbits=15)
 | |
|         zlib15 = co.compress(HAMLET_SCENE) + co.flush()
 | |
|         self.assertEqual(zlib.decompress(zlib15, 15), HAMLET_SCENE)
 | |
|         if supports_wbits_0:
 | |
|             self.assertEqual(zlib.decompress(zlib15, 0), HAMLET_SCENE)
 | |
|         self.assertEqual(zlib.decompress(zlib15, 32 + 15), HAMLET_SCENE)
 | |
|         with self.assertRaisesRegex(zlib.error, 'invalid window size'):
 | |
|             zlib.decompress(zlib15, 14)
 | |
|         dco = zlib.decompressobj(wbits=32 + 15)
 | |
|         self.assertEqual(dco.decompress(zlib15), HAMLET_SCENE)
 | |
|         dco = zlib.decompressobj(wbits=14)
 | |
|         with self.assertRaisesRegex(zlib.error, 'invalid window size'):
 | |
|             dco.decompress(zlib15)
 | |
| 
 | |
|         co = zlib.compressobj(level=1, wbits=9)
 | |
|         zlib9 = co.compress(HAMLET_SCENE) + co.flush()
 | |
|         self.assertEqual(zlib.decompress(zlib9, 9), HAMLET_SCENE)
 | |
|         self.assertEqual(zlib.decompress(zlib9, 15), HAMLET_SCENE)
 | |
|         if supports_wbits_0:
 | |
|             self.assertEqual(zlib.decompress(zlib9, 0), HAMLET_SCENE)
 | |
|         self.assertEqual(zlib.decompress(zlib9, 32 + 9), HAMLET_SCENE)
 | |
|         dco = zlib.decompressobj(wbits=32 + 9)
 | |
|         self.assertEqual(dco.decompress(zlib9), HAMLET_SCENE)
 | |
| 
 | |
|         co = zlib.compressobj(level=1, wbits=-15)
 | |
|         deflate15 = co.compress(HAMLET_SCENE) + co.flush()
 | |
|         self.assertEqual(zlib.decompress(deflate15, -15), HAMLET_SCENE)
 | |
|         dco = zlib.decompressobj(wbits=-15)
 | |
|         self.assertEqual(dco.decompress(deflate15), HAMLET_SCENE)
 | |
| 
 | |
|         co = zlib.compressobj(level=1, wbits=-9)
 | |
|         deflate9 = co.compress(HAMLET_SCENE) + co.flush()
 | |
|         self.assertEqual(zlib.decompress(deflate9, -9), HAMLET_SCENE)
 | |
|         self.assertEqual(zlib.decompress(deflate9, -15), HAMLET_SCENE)
 | |
|         dco = zlib.decompressobj(wbits=-9)
 | |
|         self.assertEqual(dco.decompress(deflate9), HAMLET_SCENE)
 | |
| 
 | |
|         co = zlib.compressobj(level=1, wbits=16 + 15)
 | |
|         gzip = co.compress(HAMLET_SCENE) + co.flush()
 | |
|         self.assertEqual(zlib.decompress(gzip, 16 + 15), HAMLET_SCENE)
 | |
|         self.assertEqual(zlib.decompress(gzip, 32 + 15), HAMLET_SCENE)
 | |
|         dco = zlib.decompressobj(32 + 15)
 | |
|         self.assertEqual(dco.decompress(gzip), HAMLET_SCENE)
 | |
| 
 | |
| 
 | |
| def genblock(seed, length, step=1024, generator=random):
 | |
|     """length-byte stream of random data from a seed (in step-byte blocks)."""
 | |
|     if seed is not None:
 | |
|         generator.seed(seed)
 | |
|     randint = generator.randint
 | |
|     if length < step or step < 2:
 | |
|         step = length
 | |
|     blocks = bytes()
 | |
|     for i in range(0, length, step):
 | |
|         blocks += bytes(randint(0, 255) for x in range(step))
 | |
|     return blocks
 | |
| 
 | |
| 
 | |
| 
 | |
| def choose_lines(source, number, seed=None, generator=random):
 | |
|     """Return a list of number lines randomly chosen from the source"""
 | |
|     if seed is not None:
 | |
|         generator.seed(seed)
 | |
|     sources = source.split('\n')
 | |
|     return [generator.choice(sources) for n in range(number)]
 | |
| 
 | |
| 
 | |
| 
 | |
| HAMLET_SCENE = b"""
 | |
| LAERTES
 | |
| 
 | |
|        O, fear me not.
 | |
|        I stay too long: but here my father comes.
 | |
| 
 | |
|        Enter POLONIUS
 | |
| 
 | |
|        A double blessing is a double grace,
 | |
|        Occasion smiles upon a second leave.
 | |
| 
 | |
| LORD POLONIUS
 | |
| 
 | |
|        Yet here, Laertes! aboard, aboard, for shame!
 | |
|        The wind sits in the shoulder of your sail,
 | |
|        And you are stay'd for. There; my blessing with thee!
 | |
|        And these few precepts in thy memory
 | |
|        See thou character. Give thy thoughts no tongue,
 | |
|        Nor any unproportioned thought his act.
 | |
|        Be thou familiar, but by no means vulgar.
 | |
|        Those friends thou hast, and their adoption tried,
 | |
|        Grapple them to thy soul with hoops of steel;
 | |
|        But do not dull thy palm with entertainment
 | |
|        Of each new-hatch'd, unfledged comrade. Beware
 | |
|        Of entrance to a quarrel, but being in,
 | |
|        Bear't that the opposed may beware of thee.
 | |
|        Give every man thy ear, but few thy voice;
 | |
|        Take each man's censure, but reserve thy judgment.
 | |
|        Costly thy habit as thy purse can buy,
 | |
|        But not express'd in fancy; rich, not gaudy;
 | |
|        For the apparel oft proclaims the man,
 | |
|        And they in France of the best rank and station
 | |
|        Are of a most select and generous chief in that.
 | |
|        Neither a borrower nor a lender be;
 | |
|        For loan oft loses both itself and friend,
 | |
|        And borrowing dulls the edge of husbandry.
 | |
|        This above all: to thine ownself be true,
 | |
|        And it must follow, as the night the day,
 | |
|        Thou canst not then be false to any man.
 | |
|        Farewell: my blessing season this in thee!
 | |
| 
 | |
| LAERTES
 | |
| 
 | |
|        Most humbly do I take my leave, my lord.
 | |
| 
 | |
| LORD POLONIUS
 | |
| 
 | |
|        The time invites you; go; your servants tend.
 | |
| 
 | |
| LAERTES
 | |
| 
 | |
|        Farewell, Ophelia; and remember well
 | |
|        What I have said to you.
 | |
| 
 | |
| OPHELIA
 | |
| 
 | |
|        'Tis in my memory lock'd,
 | |
|        And you yourself shall keep the key of it.
 | |
| 
 | |
| LAERTES
 | |
| 
 | |
|        Farewell.
 | |
| """
 | |
| 
 | |
| 
 | |
| class CustomInt:
 | |
|     def __index__(self):
 | |
|         return 100
 | |
| 
 | |
| 
 | |
| if __name__ == "__main__":
 | |
|     unittest.main()
 |