[Python-checkins] python/dist/src/Lib/test test_zlib.py, 1.23.8.1,
1.23.8.2
akuchling at users.sourceforge.net
akuchling at users.sourceforge.net
Tue Dec 28 21:26:02 CET 2004
Update of /cvsroot/python/python/dist/src/Lib/test
In directory sc8-pr-cvs1.sourceforge.net:/tmp/cvs-serv30126/Lib/test
Modified Files:
Tag: release23-maint
test_zlib.py
Log Message:
Use zlib test suite from the trunk (rev. 2.67) on the 2.3 maintenance branch
Index: test_zlib.py
===================================================================
RCS file: /cvsroot/python/python/dist/src/Lib/test/test_zlib.py,v
retrieving revision 1.23.8.1
retrieving revision 1.23.8.2
diff -u -d -r1.23.8.1 -r1.23.8.2
--- test_zlib.py 31 Aug 2003 04:37:25 -0000 1.23.8.1
+++ test_zlib.py 28 Dec 2004 20:26:00 -0000 1.23.8.2
@@ -63,7 +63,7 @@
def test_badcompressobj(self):
# verify failure on building compress object with bad params
- self.assertRaises(ValueError, zlib.compressobj, 1, 8, 0)
+ self.assertRaises(ValueError, zlib.compressobj, 1, zlib.DEFLATED, 0)
def test_baddecompressobj(self):
# verify failure on building decompress object with bad params
@@ -74,25 +74,12 @@
class CompressTestCase(unittest.TestCase):
# Test compression in one go (whole message compression)
def test_speech(self):
- # decompress(compress(data)) better be data
- x = zlib.compress(hamlet_scene)
- self.assertEqual(zlib.decompress(x), hamlet_scene)
-
- def test_speech8(self):
- # decompress(compress(data)) better be data -- more compression chances
- data = hamlet_scene * 8
- x = zlib.compress(data)
- self.assertEqual(zlib.decompress(x), data)
-
- def test_speech16(self):
- # decompress(compress(data)) better be data -- more compression chances
- data = hamlet_scene * 16
- x = zlib.compress(data)
- self.assertEqual(zlib.decompress(x), data)
+ x = zlib.compress(HAMLET_SCENE)
+ self.assertEqual(zlib.decompress(x), HAMLET_SCENE)
def test_speech128(self):
- # decompress(compress(data)) better be data -- more compression chances
- data = hamlet_scene * 8 * 16
+ # compress more data
+ data = HAMLET_SCENE * 128
x = zlib.compress(data)
self.assertEqual(zlib.decompress(x), data)
@@ -101,186 +88,126 @@
class CompressObjectTestCase(unittest.TestCase):
# Test compression object
- def test_pairsmall(self):
- # use compress object in straightforward manner, decompress w/ object
- data = hamlet_scene
- co = zlib.compressobj(8, 8, -15)
+ def test_pair(self):
+ # straightforward compress/decompress objects
+ data = HAMLET_SCENE * 128
+ co = zlib.compressobj()
x1 = co.compress(data)
x2 = co.flush()
self.assertRaises(zlib.error, co.flush) # second flush should not work
- dco = zlib.decompressobj(-15)
+ dco = zlib.decompressobj()
y1 = dco.decompress(x1 + x2)
y2 = dco.flush()
self.assertEqual(data, y1 + y2)
- def test_pair(self):
- # straightforward compress/decompress objects, more compression
- data = hamlet_scene * 8 * 16
- co = zlib.compressobj(8, 8, -15)
- x1 = co.compress(data)
+ def test_compressoptions(self):
+ # specify lots of options to compressobj()
+ level = 2
+ method = zlib.DEFLATED
+ wbits = -12
+ memlevel = 9
+ strategy = zlib.Z_FILTERED
+ co = zlib.compressobj(level, method, wbits, memlevel, strategy)
+ x1 = co.compress(HAMLET_SCENE)
x2 = co.flush()
- self.assertRaises(zlib.error, co.flush) # second flush should not work
- dco = zlib.decompressobj(-15)
+ dco = zlib.decompressobj(wbits)
y1 = dco.decompress(x1 + x2)
y2 = dco.flush()
- self.assertEqual(data, y1 + y2)
+ self.assertEqual(HAMLET_SCENE, y1 + y2)
def test_compressincremental(self):
# compress object in steps, decompress object as one-shot
- data = hamlet_scene * 8 * 16
- co = zlib.compressobj(2, 8, -12, 9, 1)
+ data = HAMLET_SCENE * 128
+ co = zlib.compressobj()
bufs = []
for i in range(0, len(data), 256):
bufs.append(co.compress(data[i:i+256]))
bufs.append(co.flush())
combuf = ''.join(bufs)
- dco = zlib.decompressobj(-15)
+ dco = zlib.decompressobj()
y1 = dco.decompress(''.join(bufs))
y2 = dco.flush()
self.assertEqual(data, y1 + y2)
- def test_decompressincremental(self):
+ def test_decompinc(self, flush=False, source=None, cx=256, dcx=64):
# compress object in steps, decompress object in steps
- data = hamlet_scene * 8 * 16
- co = zlib.compressobj(2, 8, -12, 9, 1)
+ source = source or HAMLET_SCENE
+ data = source * 128
+ co = zlib.compressobj()
bufs = []
- for i in range(0, len(data), 256):
- bufs.append(co.compress(data[i:i+256]))
+ for i in range(0, len(data), cx):
+ bufs.append(co.compress(data[i:i+cx]))
bufs.append(co.flush())
combuf = ''.join(bufs)
- self.assertEqual(data, zlib.decompress(combuf, -12, -5))
+ self.assertEqual(data, zlib.decompress(combuf))
- dco = zlib.decompressobj(-12)
+ dco = zlib.decompressobj()
bufs = []
- for i in range(0, len(combuf), 128):
- bufs.append(dco.decompress(combuf[i:i+128]))
+ for i in range(0, len(combuf), dcx):
+ bufs.append(dco.decompress(combuf[i:i+dcx]))
self.assertEqual('', dco.unconsumed_tail, ########
"(A) uct should be '': not %d long" %
- len(dco.unconsumed_tail))
- bufs.append(dco.flush())
+ len(dco.unconsumed_tail))
+ if flush:
+ bufs.append(dco.flush())
+ else:
+ while True:
+ chunk = dco.decompress('')
+ if chunk:
+ bufs.append(chunk)
+ else:
+ break
self.assertEqual('', dco.unconsumed_tail, ########
- "(B) uct should be '': not %d long" %
- len(dco.unconsumed_tail))
+ "(B) uct should be '': not %d long" %
+ len(dco.unconsumed_tail))
self.assertEqual(data, ''.join(bufs))
# Failure means: "decompressobj with init options failed"
- def test_decompinc(self,sizes=[128],flush=True,source=None,cx=256,dcx=64):
- # compress object in steps, decompress object in steps, loop sizes
- source = source or hamlet_scene
- for reps in sizes:
- data = source * reps
- co = zlib.compressobj(2, 8, -12, 9, 1)
- bufs = []
- for i in range(0, len(data), cx):
- bufs.append(co.compress(data[i:i+cx]))
- bufs.append(co.flush())
- combuf = ''.join(bufs)
-
- self.assertEqual(data, zlib.decompress(combuf, -12, -5))
-
- dco = zlib.decompressobj(-12)
- bufs = []
- for i in range(0, len(combuf), dcx):
- bufs.append(dco.decompress(combuf[i:i+dcx]))
- self.assertEqual('', dco.unconsumed_tail, ########
- "(A) uct should be '': not %d long" %
- len(dco.unconsumed_tail))
- if flush:
- bufs.append(dco.flush())
- else:
- while True:
- chunk = dco.decompress('')
- if chunk:
- bufs.append(chunk)
- else:
- break
- self.assertEqual('', dco.unconsumed_tail, ########
- "(B) uct should be '': not %d long" %
- len(dco.unconsumed_tail))
- self.assertEqual(data, ''.join(bufs))
- # Failure means: "decompressobj with init options failed"
-
- def test_decompimax(self,sizes=[128],flush=True,source=None,cx=256,dcx=64):
- # compress in steps, decompress in length-restricted steps, loop sizes
- source = source or hamlet_scene
- for reps in sizes:
- # Check a decompression object with max_length specified
- data = source * reps
- co = zlib.compressobj(2, 8, -12, 9, 1)
- bufs = []
- for i in range(0, len(data), cx):
- bufs.append(co.compress(data[i:i+cx]))
- bufs.append(co.flush())
- combuf = ''.join(bufs)
- self.assertEqual(data, zlib.decompress(combuf, -12, -5),
- 'compressed data failure')
-
- dco = zlib.decompressobj(-12)
- bufs = []
- cb = combuf
- while cb:
- #max_length = 1 + len(cb)//10
- chunk = dco.decompress(cb, dcx)
- self.failIf(len(chunk) > dcx,
- 'chunk too big (%d>%d)' % (len(chunk), dcx))
- bufs.append(chunk)
- cb = dco.unconsumed_tail
- if flush:
- bufs.append(dco.flush())
- else:
- while True:
- chunk = dco.decompress('', dcx)
- self.failIf(len(chunk) > dcx,
- 'chunk too big in tail (%d>%d)' % (len(chunk), dcx))
- if chunk:
- bufs.append(chunk)
- else:
- break
- self.assertEqual(len(data), len(''.join(bufs)))
- self.assertEqual(data, ''.join(bufs), 'Wrong data retrieved')
+ def test_decompincflush(self):
+ self.test_decompinc(flush=True)
- def test_decompressmaxlen(self):
+ def test_decompimax(self, source=None, cx=256, dcx=64):
+ # compress in steps, decompress in length-restricted steps
+ source = source or HAMLET_SCENE
# Check a decompression object with max_length specified
- data = hamlet_scene * 8 * 16
- co = zlib.compressobj(2, 8, -12, 9, 1)
+ data = source * 128
+ co = zlib.compressobj()
bufs = []
- for i in range(0, len(data), 256):
- bufs.append(co.compress(data[i:i+256]))
+ for i in range(0, len(data), cx):
+ bufs.append(co.compress(data[i:i+cx]))
bufs.append(co.flush())
combuf = ''.join(bufs)
- self.assertEqual(data, zlib.decompress(combuf, -12, -5),
+ self.assertEqual(data, zlib.decompress(combuf),
'compressed data failure')
- dco = zlib.decompressobj(-12)
+ dco = zlib.decompressobj()
bufs = []
cb = combuf
while cb:
- max_length = 1 + len(cb)//10
- chunk = dco.decompress(cb, max_length)
- self.failIf(len(chunk) > max_length,
- 'chunk too big (%d>%d)' % (len(chunk),max_length))
+ #max_length = 1 + len(cb)//10
+ chunk = dco.decompress(cb, dcx)
+ self.failIf(len(chunk) > dcx,
+ 'chunk too big (%d>%d)' % (len(chunk), dcx))
bufs.append(chunk)
cb = dco.unconsumed_tail
bufs.append(dco.flush())
- self.assertEqual(len(data), len(''.join(bufs)))
self.assertEqual(data, ''.join(bufs), 'Wrong data retrieved')
- def test_decompressmaxlenflushless(self):
- # identical to test_decompressmaxlen except flush is replaced
- # with an equivalent. This works and other fails on (eg) 2.2.2
- data = hamlet_scene * 8 * 16
- co = zlib.compressobj(2, 8, -12, 9, 1)
+ def test_decompressmaxlen(self, flush=False):
+ # Check a decompression object with max_length specified
+ data = HAMLET_SCENE * 128
+ co = zlib.compressobj()
bufs = []
for i in range(0, len(data), 256):
bufs.append(co.compress(data[i:i+256]))
bufs.append(co.flush())
combuf = ''.join(bufs)
- self.assertEqual(data, zlib.decompress(combuf, -12, -5),
- 'compressed data mismatch')
+ self.assertEqual(data, zlib.decompress(combuf),
+ 'compressed data failure')
- dco = zlib.decompressobj(-12)
+ dco = zlib.decompressobj()
bufs = []
cb = combuf
while cb:
@@ -290,19 +217,22 @@
'chunk too big (%d>%d)' % (len(chunk),max_length))
bufs.append(chunk)
cb = dco.unconsumed_tail
-
- #bufs.append(dco.flush())
- while len(chunk):
- chunk = dco.decompress('', max_length)
- self.failIf(len(chunk) > max_length,
- 'chunk too big (%d>%d)' % (len(chunk),max_length))
- bufs.append(chunk)
-
+ if flush:
+ bufs.append(dco.flush())
+ else:
+ while chunk:
+ chunk = dco.decompress('', max_length)
+ self.failIf(len(chunk) > max_length,
+ 'chunk too big (%d>%d)' % (len(chunk),max_length))
+ bufs.append(chunk)
self.assertEqual(data, ''.join(bufs), 'Wrong data retrieved')
+ def test_decompressmaxlenflush(self):
+ self.test_decompressmaxlen(flush=True)
+
def test_maxlenmisc(self):
# Misc tests of max_length
- dco = zlib.decompressobj(-12)
+ dco = zlib.decompressobj()
self.assertRaises(ValueError, dco.decompress, "", -1)
self.assertEqual('', dco.unconsumed_tail)
@@ -312,7 +242,7 @@
sync_opt = ['Z_NO_FLUSH', 'Z_SYNC_FLUSH', 'Z_FULL_FLUSH']
sync_opt = [getattr(zlib, opt) for opt in sync_opt
if hasattr(zlib, opt)]
- data = hamlet_scene * 8
+ data = HAMLET_SCENE * 8
for sync in sync_opt:
for level in range(10):
@@ -334,7 +264,7 @@
# Testing on 17K of "random" data
# Create compressor and decompressor objects
- co = zlib.compressobj(9)
+ co = zlib.compressobj(zlib.Z_BEST_COMPRESSION)
dco = zlib.decompressobj()
# Try 17K of data
@@ -360,23 +290,16 @@
# if decompressed data is different from the input data, choke.
self.assertEqual(expanded, data, "17K random source doesn't match")
- def test_manydecompinc(self):
- # Run incremental decompress test for a large range of sizes
- self.test_decompinc(sizes=[1<<n for n in range(8)],
- flush=True, cx=32, dcx=4)
-
- def test_manydecompimax(self):
- # Run incremental decompress maxlen test for a large range of sizes
- # avoid the flush bug
- self.test_decompimax(sizes=[1<<n for n in range(8)],
- flush=False, cx=32, dcx=4)
-
- def test_manydecompimaxflush(self):
- # Run incremental decompress maxlen test for a large range of sizes
- # avoid the flush bug
- self.test_decompimax(sizes=[1<<n for n in range(8)],
- flush=True, cx=32, dcx=4)
+ def test_empty_flush(self):
+ # Test that calling .flush() on unused objects works.
+ # (Bug #1083110 -- calling .flush() on decompress objects
+ # caused a core dump.)
+ co = zlib.compressobj(zlib.Z_BEST_COMPRESSION)
+ self.failUnless(co.flush()) # Returns a zlib header
+ dco = zlib.decompressobj()
+ self.assertEqual(dco.flush(), "") # Returns nothing
+
def genblock(seed, length, step=1024, generator=random):
"""length-byte stream of random data from a seed (in step-byte blocks)."""
@@ -402,7 +325,7 @@
-hamlet_scene = """
+HAMLET_SCENE = """
LAERTES
O, fear me not.
More information about the Python-checkins
mailing list