3876 lines
		
	
	
		
			137 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			3876 lines
		
	
	
		
			137 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| import collections
 | |
| import copyreg
 | |
| import dbm
 | |
| import io
 | |
| import functools
 | |
| import os
 | |
| import math
 | |
| import pickle
 | |
| import pickletools
 | |
| import shutil
 | |
| import struct
 | |
| import sys
 | |
| import threading
 | |
| import unittest
 | |
| import weakref
 | |
| from textwrap import dedent
 | |
| from http.cookies import SimpleCookie
 | |
| 
 | |
| try:
 | |
|     import _testbuffer
 | |
| except ImportError:
 | |
|     _testbuffer = None
 | |
| 
 | |
| from test import support
 | |
| from test.support import (
 | |
|     TestFailed, TESTFN, run_with_locale, no_tracing,
 | |
|     _2G, _4G, bigmemtest, reap_threads, forget,
 | |
|     save_restore_warnings_filters
 | |
|     )
 | |
| 
 | |
| from pickle import bytes_types
 | |
| 
 | |
| 
 | |
| # bpo-41003: Save/restore warnings filters to leave them unchanged.
 | |
| # Ignore filters installed by numpy.
 | |
| try:
 | |
|     with save_restore_warnings_filters():
 | |
|         import numpy as np
 | |
| except ImportError:
 | |
|     np = None
 | |
| 
 | |
| 
 | |
| requires_32b = unittest.skipUnless(sys.maxsize < 2**32,
 | |
|                                    "test is only meaningful on 32-bit builds")
 | |
| 
 | |
| # Tests that try a number of pickle protocols should have a
 | |
| #     for proto in protocols:
 | |
| # kind of outer loop.
 | |
| protocols = range(pickle.HIGHEST_PROTOCOL + 1)
 | |
| 
 | |
| 
 | |
| # Return True if opcode code appears in the pickle, else False.
 | |
| def opcode_in_pickle(code, pickle):
 | |
|     for op, dummy, dummy in pickletools.genops(pickle):
 | |
|         if op.code == code.decode("latin-1"):
 | |
|             return True
 | |
|     return False
 | |
| 
 | |
| # Return the number of times opcode code appears in pickle.
 | |
| def count_opcode(code, pickle):
 | |
|     n = 0
 | |
|     for op, dummy, dummy in pickletools.genops(pickle):
 | |
|         if op.code == code.decode("latin-1"):
 | |
|             n += 1
 | |
|     return n
 | |
| 
 | |
| 
 | |
| def identity(x):
 | |
|     return x
 | |
| 
 | |
| 
 | |
| class UnseekableIO(io.BytesIO):
 | |
|     def peek(self, *args):
 | |
|         raise NotImplementedError
 | |
| 
 | |
|     def seekable(self):
 | |
|         return False
 | |
| 
 | |
|     def seek(self, *args):
 | |
|         raise io.UnsupportedOperation
 | |
| 
 | |
|     def tell(self):
 | |
|         raise io.UnsupportedOperation
 | |
| 
 | |
| 
 | |
| class MinimalIO(object):
 | |
|     """
 | |
|     A file-like object that doesn't support readinto().
 | |
|     """
 | |
|     def __init__(self, *args):
 | |
|         self._bio = io.BytesIO(*args)
 | |
|         self.getvalue = self._bio.getvalue
 | |
|         self.read = self._bio.read
 | |
|         self.readline = self._bio.readline
 | |
|         self.write = self._bio.write
 | |
| 
 | |
| 
 | |
| # We can't very well test the extension registry without putting known stuff
 | |
| # in it, but we have to be careful to restore its original state.  Code
 | |
| # should do this:
 | |
| #
 | |
| #     e = ExtensionSaver(extension_code)
 | |
| #     try:
 | |
| #         fiddle w/ the extension registry's stuff for extension_code
 | |
| #     finally:
 | |
| #         e.restore()
 | |
| 
 | |
| class ExtensionSaver:
 | |
|     # Remember current registration for code (if any), and remove it (if
 | |
|     # there is one).
 | |
|     def __init__(self, code):
 | |
|         self.code = code
 | |
|         if code in copyreg._inverted_registry:
 | |
|             self.pair = copyreg._inverted_registry[code]
 | |
|             copyreg.remove_extension(self.pair[0], self.pair[1], code)
 | |
|         else:
 | |
|             self.pair = None
 | |
| 
 | |
|     # Restore previous registration for code.
 | |
|     def restore(self):
 | |
|         code = self.code
 | |
|         curpair = copyreg._inverted_registry.get(code)
 | |
|         if curpair is not None:
 | |
|             copyreg.remove_extension(curpair[0], curpair[1], code)
 | |
|         pair = self.pair
 | |
|         if pair is not None:
 | |
|             copyreg.add_extension(pair[0], pair[1], code)
 | |
| 
 | |
| class C:
 | |
|     def __eq__(self, other):
 | |
|         return self.__dict__ == other.__dict__
 | |
| 
 | |
| class D(C):
 | |
|     def __init__(self, arg):
 | |
|         pass
 | |
| 
 | |
| class E(C):
 | |
|     def __getinitargs__(self):
 | |
|         return ()
 | |
| 
 | |
| # Simple mutable object.
 | |
| class Object:
 | |
|     pass
 | |
| 
 | |
| # Hashable immutable key object containing unheshable mutable data.
 | |
| class K:
 | |
|     def __init__(self, value):
 | |
|         self.value = value
 | |
| 
 | |
|     def __reduce__(self):
 | |
|         # Shouldn't support the recursion itself
 | |
|         return K, (self.value,)
 | |
| 
 | |
| import __main__
 | |
| __main__.C = C
 | |
| C.__module__ = "__main__"
 | |
| __main__.D = D
 | |
| D.__module__ = "__main__"
 | |
| __main__.E = E
 | |
| E.__module__ = "__main__"
 | |
| 
 | |
| class myint(int):
 | |
|     def __init__(self, x):
 | |
|         self.str = str(x)
 | |
| 
 | |
| class initarg(C):
 | |
| 
 | |
|     def __init__(self, a, b):
 | |
|         self.a = a
 | |
|         self.b = b
 | |
| 
 | |
|     def __getinitargs__(self):
 | |
|         return self.a, self.b
 | |
| 
 | |
| class metaclass(type):
 | |
|     pass
 | |
| 
 | |
| class use_metaclass(object, metaclass=metaclass):
 | |
|     pass
 | |
| 
 | |
| class pickling_metaclass(type):
 | |
|     def __eq__(self, other):
 | |
|         return (type(self) == type(other) and
 | |
|                 self.reduce_args == other.reduce_args)
 | |
| 
 | |
|     def __reduce__(self):
 | |
|         return (create_dynamic_class, self.reduce_args)
 | |
| 
 | |
| def create_dynamic_class(name, bases):
 | |
|     result = pickling_metaclass(name, bases, dict())
 | |
|     result.reduce_args = (name, bases)
 | |
|     return result
 | |
| 
 | |
| 
 | |
| class ZeroCopyBytes(bytes):
 | |
|     readonly = True
 | |
|     c_contiguous = True
 | |
|     f_contiguous = True
 | |
|     zero_copy_reconstruct = True
 | |
| 
 | |
|     def __reduce_ex__(self, protocol):
 | |
|         if protocol >= 5:
 | |
|             return type(self)._reconstruct, (pickle.PickleBuffer(self),), None
 | |
|         else:
 | |
|             return type(self)._reconstruct, (bytes(self),)
 | |
| 
 | |
|     def __repr__(self):
 | |
|         return "{}({!r})".format(self.__class__.__name__, bytes(self))
 | |
| 
 | |
|     __str__ = __repr__
 | |
| 
 | |
|     @classmethod
 | |
|     def _reconstruct(cls, obj):
 | |
|         with memoryview(obj) as m:
 | |
|             obj = m.obj
 | |
|             if type(obj) is cls:
 | |
|                 # Zero-copy
 | |
|                 return obj
 | |
|             else:
 | |
|                 return cls(obj)
 | |
| 
 | |
| 
 | |
| class ZeroCopyBytearray(bytearray):
 | |
|     readonly = False
 | |
|     c_contiguous = True
 | |
|     f_contiguous = True
 | |
|     zero_copy_reconstruct = True
 | |
| 
 | |
|     def __reduce_ex__(self, protocol):
 | |
|         if protocol >= 5:
 | |
|             return type(self)._reconstruct, (pickle.PickleBuffer(self),), None
 | |
|         else:
 | |
|             return type(self)._reconstruct, (bytes(self),)
 | |
| 
 | |
|     def __repr__(self):
 | |
|         return "{}({!r})".format(self.__class__.__name__, bytes(self))
 | |
| 
 | |
|     __str__ = __repr__
 | |
| 
 | |
|     @classmethod
 | |
|     def _reconstruct(cls, obj):
 | |
|         with memoryview(obj) as m:
 | |
|             obj = m.obj
 | |
|             if type(obj) is cls:
 | |
|                 # Zero-copy
 | |
|                 return obj
 | |
|             else:
 | |
|                 return cls(obj)
 | |
| 
 | |
| 
 | |
| if _testbuffer is not None:
 | |
| 
 | |
|     class PicklableNDArray:
 | |
|         # A not-really-zero-copy picklable ndarray, as the ndarray()
 | |
|         # constructor doesn't allow for it
 | |
| 
 | |
|         zero_copy_reconstruct = False
 | |
| 
 | |
|         def __init__(self, *args, **kwargs):
 | |
|             self.array = _testbuffer.ndarray(*args, **kwargs)
 | |
| 
 | |
|         def __getitem__(self, idx):
 | |
|             cls = type(self)
 | |
|             new = cls.__new__(cls)
 | |
|             new.array = self.array[idx]
 | |
|             return new
 | |
| 
 | |
|         @property
 | |
|         def readonly(self):
 | |
|             return self.array.readonly
 | |
| 
 | |
|         @property
 | |
|         def c_contiguous(self):
 | |
|             return self.array.c_contiguous
 | |
| 
 | |
|         @property
 | |
|         def f_contiguous(self):
 | |
|             return self.array.f_contiguous
 | |
| 
 | |
|         def __eq__(self, other):
 | |
|             if not isinstance(other, PicklableNDArray):
 | |
|                 return NotImplemented
 | |
|             return (other.array.format == self.array.format and
 | |
|                     other.array.shape == self.array.shape and
 | |
|                     other.array.strides == self.array.strides and
 | |
|                     other.array.readonly == self.array.readonly and
 | |
|                     other.array.tobytes() == self.array.tobytes())
 | |
| 
 | |
|         def __ne__(self, other):
 | |
|             if not isinstance(other, PicklableNDArray):
 | |
|                 return NotImplemented
 | |
|             return not (self == other)
 | |
| 
 | |
|         def __repr__(self):
 | |
|             return (f"{type(self)}(shape={self.array.shape},"
 | |
|                     f"strides={self.array.strides}, "
 | |
|                     f"bytes={self.array.tobytes()})")
 | |
| 
 | |
|         def __reduce_ex__(self, protocol):
 | |
|             if not self.array.contiguous:
 | |
|                 raise NotImplementedError("Reconstructing a non-contiguous "
 | |
|                                           "ndarray does not seem possible")
 | |
|             ndarray_kwargs = {"shape": self.array.shape,
 | |
|                               "strides": self.array.strides,
 | |
|                               "format": self.array.format,
 | |
|                               "flags": (0 if self.readonly
 | |
|                                         else _testbuffer.ND_WRITABLE)}
 | |
|             pb = pickle.PickleBuffer(self.array)
 | |
|             if protocol >= 5:
 | |
|                 return (type(self)._reconstruct,
 | |
|                         (pb, ndarray_kwargs))
 | |
|             else:
 | |
|                 # Need to serialize the bytes in physical order
 | |
|                 with pb.raw() as m:
 | |
|                     return (type(self)._reconstruct,
 | |
|                             (m.tobytes(), ndarray_kwargs))
 | |
| 
 | |
|         @classmethod
 | |
|         def _reconstruct(cls, obj, kwargs):
 | |
|             with memoryview(obj) as m:
 | |
|                 # For some reason, ndarray() wants a list of integers...
 | |
|                 # XXX This only works if format == 'B'
 | |
|                 items = list(m.tobytes())
 | |
|             return cls(items, **kwargs)
 | |
| 
 | |
| 
 | |
| # DATA0 .. DATA4 are the pickles we expect under the various protocols, for
 | |
| # the object returned by create_data().
 | |
| 
 | |
| DATA0 = (
 | |
|     b'(lp0\nL0L\naL1L\naF2.0\n'
 | |
|     b'ac__builtin__\ncomple'
 | |
|     b'x\np1\n(F3.0\nF0.0\ntp2\n'
 | |
|     b'Rp3\naL1L\naL-1L\naL255'
 | |
|     b'L\naL-255L\naL-256L\naL'
 | |
|     b'65535L\naL-65535L\naL-'
 | |
|     b'65536L\naL2147483647L'
 | |
|     b'\naL-2147483647L\naL-2'
 | |
|     b'147483648L\na(Vabc\np4'
 | |
|     b'\ng4\nccopy_reg\n_recon'
 | |
|     b'structor\np5\n(c__main'
 | |
|     b'__\nC\np6\nc__builtin__'
 | |
|     b'\nobject\np7\nNtp8\nRp9\n'
 | |
|     b'(dp10\nVfoo\np11\nL1L\ns'
 | |
|     b'Vbar\np12\nL2L\nsbg9\ntp'
 | |
|     b'13\nag13\naL5L\na.'
 | |
| )
 | |
| 
 | |
| # Disassembly of DATA0
 | |
| DATA0_DIS = """\
 | |
|     0: (    MARK
 | |
|     1: l        LIST       (MARK at 0)
 | |
|     2: p    PUT        0
 | |
|     5: L    LONG       0
 | |
|     9: a    APPEND
 | |
|    10: L    LONG       1
 | |
|    14: a    APPEND
 | |
|    15: F    FLOAT      2.0
 | |
|    20: a    APPEND
 | |
|    21: c    GLOBAL     '__builtin__ complex'
 | |
|    42: p    PUT        1
 | |
|    45: (    MARK
 | |
|    46: F        FLOAT      3.0
 | |
|    51: F        FLOAT      0.0
 | |
|    56: t        TUPLE      (MARK at 45)
 | |
|    57: p    PUT        2
 | |
|    60: R    REDUCE
 | |
|    61: p    PUT        3
 | |
|    64: a    APPEND
 | |
|    65: L    LONG       1
 | |
|    69: a    APPEND
 | |
|    70: L    LONG       -1
 | |
|    75: a    APPEND
 | |
|    76: L    LONG       255
 | |
|    82: a    APPEND
 | |
|    83: L    LONG       -255
 | |
|    90: a    APPEND
 | |
|    91: L    LONG       -256
 | |
|    98: a    APPEND
 | |
|    99: L    LONG       65535
 | |
|   107: a    APPEND
 | |
|   108: L    LONG       -65535
 | |
|   117: a    APPEND
 | |
|   118: L    LONG       -65536
 | |
|   127: a    APPEND
 | |
|   128: L    LONG       2147483647
 | |
|   141: a    APPEND
 | |
|   142: L    LONG       -2147483647
 | |
|   156: a    APPEND
 | |
|   157: L    LONG       -2147483648
 | |
|   171: a    APPEND
 | |
|   172: (    MARK
 | |
|   173: V        UNICODE    'abc'
 | |
|   178: p        PUT        4
 | |
|   181: g        GET        4
 | |
|   184: c        GLOBAL     'copy_reg _reconstructor'
 | |
|   209: p        PUT        5
 | |
|   212: (        MARK
 | |
|   213: c            GLOBAL     '__main__ C'
 | |
|   225: p            PUT        6
 | |
|   228: c            GLOBAL     '__builtin__ object'
 | |
|   248: p            PUT        7
 | |
|   251: N            NONE
 | |
|   252: t            TUPLE      (MARK at 212)
 | |
|   253: p        PUT        8
 | |
|   256: R        REDUCE
 | |
|   257: p        PUT        9
 | |
|   260: (        MARK
 | |
|   261: d            DICT       (MARK at 260)
 | |
|   262: p        PUT        10
 | |
|   266: V        UNICODE    'foo'
 | |
|   271: p        PUT        11
 | |
|   275: L        LONG       1
 | |
|   279: s        SETITEM
 | |
|   280: V        UNICODE    'bar'
 | |
|   285: p        PUT        12
 | |
|   289: L        LONG       2
 | |
|   293: s        SETITEM
 | |
|   294: b        BUILD
 | |
|   295: g        GET        9
 | |
|   298: t        TUPLE      (MARK at 172)
 | |
|   299: p    PUT        13
 | |
|   303: a    APPEND
 | |
|   304: g    GET        13
 | |
|   308: a    APPEND
 | |
|   309: L    LONG       5
 | |
|   313: a    APPEND
 | |
|   314: .    STOP
 | |
| highest protocol among opcodes = 0
 | |
| """
 | |
| 
 | |
| DATA1 = (
 | |
|     b']q\x00(K\x00K\x01G@\x00\x00\x00\x00\x00\x00\x00c__'
 | |
|     b'builtin__\ncomplex\nq\x01'
 | |
|     b'(G@\x08\x00\x00\x00\x00\x00\x00G\x00\x00\x00\x00\x00\x00\x00\x00t'
 | |
|     b'q\x02Rq\x03K\x01J\xff\xff\xff\xffK\xffJ\x01\xff\xff\xffJ'
 | |
|     b'\x00\xff\xff\xffM\xff\xffJ\x01\x00\xff\xffJ\x00\x00\xff\xffJ\xff\xff'
 | |
|     b'\xff\x7fJ\x01\x00\x00\x80J\x00\x00\x00\x80(X\x03\x00\x00\x00ab'
 | |
|     b'cq\x04h\x04ccopy_reg\n_reco'
 | |
|     b'nstructor\nq\x05(c__main'
 | |
|     b'__\nC\nq\x06c__builtin__\n'
 | |
|     b'object\nq\x07Ntq\x08Rq\t}q\n('
 | |
|     b'X\x03\x00\x00\x00fooq\x0bK\x01X\x03\x00\x00\x00bar'
 | |
|     b'q\x0cK\x02ubh\ttq\rh\rK\x05e.'
 | |
| )
 | |
| 
 | |
| # Disassembly of DATA1
 | |
| DATA1_DIS = """\
 | |
|     0: ]    EMPTY_LIST
 | |
|     1: q    BINPUT     0
 | |
|     3: (    MARK
 | |
|     4: K        BININT1    0
 | |
|     6: K        BININT1    1
 | |
|     8: G        BINFLOAT   2.0
 | |
|    17: c        GLOBAL     '__builtin__ complex'
 | |
|    38: q        BINPUT     1
 | |
|    40: (        MARK
 | |
|    41: G            BINFLOAT   3.0
 | |
|    50: G            BINFLOAT   0.0
 | |
|    59: t            TUPLE      (MARK at 40)
 | |
|    60: q        BINPUT     2
 | |
|    62: R        REDUCE
 | |
|    63: q        BINPUT     3
 | |
|    65: K        BININT1    1
 | |
|    67: J        BININT     -1
 | |
|    72: K        BININT1    255
 | |
|    74: J        BININT     -255
 | |
|    79: J        BININT     -256
 | |
|    84: M        BININT2    65535
 | |
|    87: J        BININT     -65535
 | |
|    92: J        BININT     -65536
 | |
|    97: J        BININT     2147483647
 | |
|   102: J        BININT     -2147483647
 | |
|   107: J        BININT     -2147483648
 | |
|   112: (        MARK
 | |
|   113: X            BINUNICODE 'abc'
 | |
|   121: q            BINPUT     4
 | |
|   123: h            BINGET     4
 | |
|   125: c            GLOBAL     'copy_reg _reconstructor'
 | |
|   150: q            BINPUT     5
 | |
|   152: (            MARK
 | |
|   153: c                GLOBAL     '__main__ C'
 | |
|   165: q                BINPUT     6
 | |
|   167: c                GLOBAL     '__builtin__ object'
 | |
|   187: q                BINPUT     7
 | |
|   189: N                NONE
 | |
|   190: t                TUPLE      (MARK at 152)
 | |
|   191: q            BINPUT     8
 | |
|   193: R            REDUCE
 | |
|   194: q            BINPUT     9
 | |
|   196: }            EMPTY_DICT
 | |
|   197: q            BINPUT     10
 | |
|   199: (            MARK
 | |
|   200: X                BINUNICODE 'foo'
 | |
|   208: q                BINPUT     11
 | |
|   210: K                BININT1    1
 | |
|   212: X                BINUNICODE 'bar'
 | |
|   220: q                BINPUT     12
 | |
|   222: K                BININT1    2
 | |
|   224: u                SETITEMS   (MARK at 199)
 | |
|   225: b            BUILD
 | |
|   226: h            BINGET     9
 | |
|   228: t            TUPLE      (MARK at 112)
 | |
|   229: q        BINPUT     13
 | |
|   231: h        BINGET     13
 | |
|   233: K        BININT1    5
 | |
|   235: e        APPENDS    (MARK at 3)
 | |
|   236: .    STOP
 | |
| highest protocol among opcodes = 1
 | |
| """
 | |
| 
 | |
| DATA2 = (
 | |
|     b'\x80\x02]q\x00(K\x00K\x01G@\x00\x00\x00\x00\x00\x00\x00c'
 | |
|     b'__builtin__\ncomplex\n'
 | |
|     b'q\x01G@\x08\x00\x00\x00\x00\x00\x00G\x00\x00\x00\x00\x00\x00\x00\x00'
 | |
|     b'\x86q\x02Rq\x03K\x01J\xff\xff\xff\xffK\xffJ\x01\xff\xff\xff'
 | |
|     b'J\x00\xff\xff\xffM\xff\xffJ\x01\x00\xff\xffJ\x00\x00\xff\xffJ\xff'
 | |
|     b'\xff\xff\x7fJ\x01\x00\x00\x80J\x00\x00\x00\x80(X\x03\x00\x00\x00a'
 | |
|     b'bcq\x04h\x04c__main__\nC\nq\x05'
 | |
|     b')\x81q\x06}q\x07(X\x03\x00\x00\x00fooq\x08K\x01'
 | |
|     b'X\x03\x00\x00\x00barq\tK\x02ubh\x06tq\nh'
 | |
|     b'\nK\x05e.'
 | |
| )
 | |
| 
 | |
| # Disassembly of DATA2
 | |
| DATA2_DIS = """\
 | |
|     0: \x80 PROTO      2
 | |
|     2: ]    EMPTY_LIST
 | |
|     3: q    BINPUT     0
 | |
|     5: (    MARK
 | |
|     6: K        BININT1    0
 | |
|     8: K        BININT1    1
 | |
|    10: G        BINFLOAT   2.0
 | |
|    19: c        GLOBAL     '__builtin__ complex'
 | |
|    40: q        BINPUT     1
 | |
|    42: G        BINFLOAT   3.0
 | |
|    51: G        BINFLOAT   0.0
 | |
|    60: \x86     TUPLE2
 | |
|    61: q        BINPUT     2
 | |
|    63: R        REDUCE
 | |
|    64: q        BINPUT     3
 | |
|    66: K        BININT1    1
 | |
|    68: J        BININT     -1
 | |
|    73: K        BININT1    255
 | |
|    75: J        BININT     -255
 | |
|    80: J        BININT     -256
 | |
|    85: M        BININT2    65535
 | |
|    88: J        BININT     -65535
 | |
|    93: J        BININT     -65536
 | |
|    98: J        BININT     2147483647
 | |
|   103: J        BININT     -2147483647
 | |
|   108: J        BININT     -2147483648
 | |
|   113: (        MARK
 | |
|   114: X            BINUNICODE 'abc'
 | |
|   122: q            BINPUT     4
 | |
|   124: h            BINGET     4
 | |
|   126: c            GLOBAL     '__main__ C'
 | |
|   138: q            BINPUT     5
 | |
|   140: )            EMPTY_TUPLE
 | |
|   141: \x81         NEWOBJ
 | |
|   142: q            BINPUT     6
 | |
|   144: }            EMPTY_DICT
 | |
|   145: q            BINPUT     7
 | |
|   147: (            MARK
 | |
|   148: X                BINUNICODE 'foo'
 | |
|   156: q                BINPUT     8
 | |
|   158: K                BININT1    1
 | |
|   160: X                BINUNICODE 'bar'
 | |
|   168: q                BINPUT     9
 | |
|   170: K                BININT1    2
 | |
|   172: u                SETITEMS   (MARK at 147)
 | |
|   173: b            BUILD
 | |
|   174: h            BINGET     6
 | |
|   176: t            TUPLE      (MARK at 113)
 | |
|   177: q        BINPUT     10
 | |
|   179: h        BINGET     10
 | |
|   181: K        BININT1    5
 | |
|   183: e        APPENDS    (MARK at 5)
 | |
|   184: .    STOP
 | |
| highest protocol among opcodes = 2
 | |
| """
 | |
| 
 | |
| DATA3 = (
 | |
|     b'\x80\x03]q\x00(K\x00K\x01G@\x00\x00\x00\x00\x00\x00\x00c'
 | |
|     b'builtins\ncomplex\nq\x01G'
 | |
|     b'@\x08\x00\x00\x00\x00\x00\x00G\x00\x00\x00\x00\x00\x00\x00\x00\x86q\x02'
 | |
|     b'Rq\x03K\x01J\xff\xff\xff\xffK\xffJ\x01\xff\xff\xffJ\x00\xff'
 | |
|     b'\xff\xffM\xff\xffJ\x01\x00\xff\xffJ\x00\x00\xff\xffJ\xff\xff\xff\x7f'
 | |
|     b'J\x01\x00\x00\x80J\x00\x00\x00\x80(X\x03\x00\x00\x00abcq'
 | |
|     b'\x04h\x04c__main__\nC\nq\x05)\x81q'
 | |
|     b'\x06}q\x07(X\x03\x00\x00\x00barq\x08K\x02X\x03\x00'
 | |
|     b'\x00\x00fooq\tK\x01ubh\x06tq\nh\nK\x05'
 | |
|     b'e.'
 | |
| )
 | |
| 
 | |
| # Disassembly of DATA3
 | |
| DATA3_DIS = """\
 | |
|     0: \x80 PROTO      3
 | |
|     2: ]    EMPTY_LIST
 | |
|     3: q    BINPUT     0
 | |
|     5: (    MARK
 | |
|     6: K        BININT1    0
 | |
|     8: K        BININT1    1
 | |
|    10: G        BINFLOAT   2.0
 | |
|    19: c        GLOBAL     'builtins complex'
 | |
|    37: q        BINPUT     1
 | |
|    39: G        BINFLOAT   3.0
 | |
|    48: G        BINFLOAT   0.0
 | |
|    57: \x86     TUPLE2
 | |
|    58: q        BINPUT     2
 | |
|    60: R        REDUCE
 | |
|    61: q        BINPUT     3
 | |
|    63: K        BININT1    1
 | |
|    65: J        BININT     -1
 | |
|    70: K        BININT1    255
 | |
|    72: J        BININT     -255
 | |
|    77: J        BININT     -256
 | |
|    82: M        BININT2    65535
 | |
|    85: J        BININT     -65535
 | |
|    90: J        BININT     -65536
 | |
|    95: J        BININT     2147483647
 | |
|   100: J        BININT     -2147483647
 | |
|   105: J        BININT     -2147483648
 | |
|   110: (        MARK
 | |
|   111: X            BINUNICODE 'abc'
 | |
|   119: q            BINPUT     4
 | |
|   121: h            BINGET     4
 | |
|   123: c            GLOBAL     '__main__ C'
 | |
|   135: q            BINPUT     5
 | |
|   137: )            EMPTY_TUPLE
 | |
|   138: \x81         NEWOBJ
 | |
|   139: q            BINPUT     6
 | |
|   141: }            EMPTY_DICT
 | |
|   142: q            BINPUT     7
 | |
|   144: (            MARK
 | |
|   145: X                BINUNICODE 'bar'
 | |
|   153: q                BINPUT     8
 | |
|   155: K                BININT1    2
 | |
|   157: X                BINUNICODE 'foo'
 | |
|   165: q                BINPUT     9
 | |
|   167: K                BININT1    1
 | |
|   169: u                SETITEMS   (MARK at 144)
 | |
|   170: b            BUILD
 | |
|   171: h            BINGET     6
 | |
|   173: t            TUPLE      (MARK at 110)
 | |
|   174: q        BINPUT     10
 | |
|   176: h        BINGET     10
 | |
|   178: K        BININT1    5
 | |
|   180: e        APPENDS    (MARK at 5)
 | |
|   181: .    STOP
 | |
| highest protocol among opcodes = 2
 | |
| """
 | |
| 
 | |
| DATA4 = (
 | |
|     b'\x80\x04\x95\xa8\x00\x00\x00\x00\x00\x00\x00]\x94(K\x00K\x01G@'
 | |
|     b'\x00\x00\x00\x00\x00\x00\x00\x8c\x08builtins\x94\x8c\x07'
 | |
|     b'complex\x94\x93\x94G@\x08\x00\x00\x00\x00\x00\x00G'
 | |
|     b'\x00\x00\x00\x00\x00\x00\x00\x00\x86\x94R\x94K\x01J\xff\xff\xff\xffK'
 | |
|     b'\xffJ\x01\xff\xff\xffJ\x00\xff\xff\xffM\xff\xffJ\x01\x00\xff\xffJ'
 | |
|     b'\x00\x00\xff\xffJ\xff\xff\xff\x7fJ\x01\x00\x00\x80J\x00\x00\x00\x80('
 | |
|     b'\x8c\x03abc\x94h\x06\x8c\x08__main__\x94\x8c'
 | |
|     b'\x01C\x94\x93\x94)\x81\x94}\x94(\x8c\x03bar\x94K\x02\x8c'
 | |
|     b'\x03foo\x94K\x01ubh\nt\x94h\x0eK\x05e.'
 | |
| )
 | |
| 
 | |
| # Disassembly of DATA4
 | |
| DATA4_DIS = """\
 | |
|     0: \x80 PROTO      4
 | |
|     2: \x95 FRAME      168
 | |
|    11: ]    EMPTY_LIST
 | |
|    12: \x94 MEMOIZE
 | |
|    13: (    MARK
 | |
|    14: K        BININT1    0
 | |
|    16: K        BININT1    1
 | |
|    18: G        BINFLOAT   2.0
 | |
|    27: \x8c     SHORT_BINUNICODE 'builtins'
 | |
|    37: \x94     MEMOIZE
 | |
|    38: \x8c     SHORT_BINUNICODE 'complex'
 | |
|    47: \x94     MEMOIZE
 | |
|    48: \x93     STACK_GLOBAL
 | |
|    49: \x94     MEMOIZE
 | |
|    50: G        BINFLOAT   3.0
 | |
|    59: G        BINFLOAT   0.0
 | |
|    68: \x86     TUPLE2
 | |
|    69: \x94     MEMOIZE
 | |
|    70: R        REDUCE
 | |
|    71: \x94     MEMOIZE
 | |
|    72: K        BININT1    1
 | |
|    74: J        BININT     -1
 | |
|    79: K        BININT1    255
 | |
|    81: J        BININT     -255
 | |
|    86: J        BININT     -256
 | |
|    91: M        BININT2    65535
 | |
|    94: J        BININT     -65535
 | |
|    99: J        BININT     -65536
 | |
|   104: J        BININT     2147483647
 | |
|   109: J        BININT     -2147483647
 | |
|   114: J        BININT     -2147483648
 | |
|   119: (        MARK
 | |
|   120: \x8c         SHORT_BINUNICODE 'abc'
 | |
|   125: \x94         MEMOIZE
 | |
|   126: h            BINGET     6
 | |
|   128: \x8c         SHORT_BINUNICODE '__main__'
 | |
|   138: \x94         MEMOIZE
 | |
|   139: \x8c         SHORT_BINUNICODE 'C'
 | |
|   142: \x94         MEMOIZE
 | |
|   143: \x93         STACK_GLOBAL
 | |
|   144: \x94         MEMOIZE
 | |
|   145: )            EMPTY_TUPLE
 | |
|   146: \x81         NEWOBJ
 | |
|   147: \x94         MEMOIZE
 | |
|   148: }            EMPTY_DICT
 | |
|   149: \x94         MEMOIZE
 | |
|   150: (            MARK
 | |
|   151: \x8c             SHORT_BINUNICODE 'bar'
 | |
|   156: \x94             MEMOIZE
 | |
|   157: K                BININT1    2
 | |
|   159: \x8c             SHORT_BINUNICODE 'foo'
 | |
|   164: \x94             MEMOIZE
 | |
|   165: K                BININT1    1
 | |
|   167: u                SETITEMS   (MARK at 150)
 | |
|   168: b            BUILD
 | |
|   169: h            BINGET     10
 | |
|   171: t            TUPLE      (MARK at 119)
 | |
|   172: \x94     MEMOIZE
 | |
|   173: h        BINGET     14
 | |
|   175: K        BININT1    5
 | |
|   177: e        APPENDS    (MARK at 13)
 | |
|   178: .    STOP
 | |
| highest protocol among opcodes = 4
 | |
| """
 | |
| 
 | |
| # set([1,2]) pickled from 2.x with protocol 2
 | |
| DATA_SET = b'\x80\x02c__builtin__\nset\nq\x00]q\x01(K\x01K\x02e\x85q\x02Rq\x03.'
 | |
| 
 | |
| # xrange(5) pickled from 2.x with protocol 2
 | |
| DATA_XRANGE = b'\x80\x02c__builtin__\nxrange\nq\x00K\x00K\x05K\x01\x87q\x01Rq\x02.'
 | |
| 
 | |
| # a SimpleCookie() object pickled from 2.x with protocol 2
 | |
| DATA_COOKIE = (b'\x80\x02cCookie\nSimpleCookie\nq\x00)\x81q\x01U\x03key'
 | |
|                b'q\x02cCookie\nMorsel\nq\x03)\x81q\x04(U\x07commentq\x05U'
 | |
|                b'\x00q\x06U\x06domainq\x07h\x06U\x06secureq\x08h\x06U\x07'
 | |
|                b'expiresq\th\x06U\x07max-ageq\nh\x06U\x07versionq\x0bh\x06U'
 | |
|                b'\x04pathq\x0ch\x06U\x08httponlyq\rh\x06u}q\x0e(U\x0b'
 | |
|                b'coded_valueq\x0fU\x05valueq\x10h\x10h\x10h\x02h\x02ubs}q\x11b.')
 | |
| 
 | |
| # set([3]) pickled from 2.x with protocol 2
 | |
| DATA_SET2 = b'\x80\x02c__builtin__\nset\nq\x00]q\x01K\x03a\x85q\x02Rq\x03.'
 | |
| 
 | |
| python2_exceptions_without_args = (
 | |
|     ArithmeticError,
 | |
|     AssertionError,
 | |
|     AttributeError,
 | |
|     BaseException,
 | |
|     BufferError,
 | |
|     BytesWarning,
 | |
|     DeprecationWarning,
 | |
|     EOFError,
 | |
|     EnvironmentError,
 | |
|     Exception,
 | |
|     FloatingPointError,
 | |
|     FutureWarning,
 | |
|     GeneratorExit,
 | |
|     IOError,
 | |
|     ImportError,
 | |
|     ImportWarning,
 | |
|     IndentationError,
 | |
|     IndexError,
 | |
|     KeyError,
 | |
|     KeyboardInterrupt,
 | |
|     LookupError,
 | |
|     MemoryError,
 | |
|     NameError,
 | |
|     NotImplementedError,
 | |
|     OSError,
 | |
|     OverflowError,
 | |
|     PendingDeprecationWarning,
 | |
|     ReferenceError,
 | |
|     RuntimeError,
 | |
|     RuntimeWarning,
 | |
|     # StandardError is gone in Python 3, we map it to Exception
 | |
|     StopIteration,
 | |
|     SyntaxError,
 | |
|     SyntaxWarning,
 | |
|     SystemError,
 | |
|     SystemExit,
 | |
|     TabError,
 | |
|     TypeError,
 | |
|     UnboundLocalError,
 | |
|     UnicodeError,
 | |
|     UnicodeWarning,
 | |
|     UserWarning,
 | |
|     ValueError,
 | |
|     Warning,
 | |
|     ZeroDivisionError,
 | |
| )
 | |
| 
 | |
| exception_pickle = b'\x80\x02cexceptions\n?\nq\x00)Rq\x01.'
 | |
| 
 | |
| # UnicodeEncodeError object pickled from 2.x with protocol 2
 | |
| DATA_UEERR = (b'\x80\x02cexceptions\nUnicodeEncodeError\n'
 | |
|               b'q\x00(U\x05asciiq\x01X\x03\x00\x00\x00fooq\x02K\x00K\x01'
 | |
|               b'U\x03badq\x03tq\x04Rq\x05.')
 | |
| 
 | |
| 
 | |
| def create_data():
 | |
|     c = C()
 | |
|     c.foo = 1
 | |
|     c.bar = 2
 | |
|     x = [0, 1, 2.0, 3.0+0j]
 | |
|     # Append some integer test cases at cPickle.c's internal size
 | |
|     # cutoffs.
 | |
|     uint1max = 0xff
 | |
|     uint2max = 0xffff
 | |
|     int4max = 0x7fffffff
 | |
|     x.extend([1, -1,
 | |
|               uint1max, -uint1max, -uint1max-1,
 | |
|               uint2max, -uint2max, -uint2max-1,
 | |
|                int4max,  -int4max,  -int4max-1])
 | |
|     y = ('abc', 'abc', c, c)
 | |
|     x.append(y)
 | |
|     x.append(y)
 | |
|     x.append(5)
 | |
|     return x
 | |
| 
 | |
| 
 | |
| class AbstractUnpickleTests(unittest.TestCase):
 | |
|     # Subclass must define self.loads.
 | |
| 
 | |
|     _testdata = create_data()
 | |
| 
 | |
|     def assert_is_copy(self, obj, objcopy, msg=None):
 | |
|         """Utility method to verify if two objects are copies of each others.
 | |
|         """
 | |
|         if msg is None:
 | |
|             msg = "{!r} is not a copy of {!r}".format(obj, objcopy)
 | |
|         self.assertEqual(obj, objcopy, msg=msg)
 | |
|         self.assertIs(type(obj), type(objcopy), msg=msg)
 | |
|         if hasattr(obj, '__dict__'):
 | |
|             self.assertDictEqual(obj.__dict__, objcopy.__dict__, msg=msg)
 | |
|             self.assertIsNot(obj.__dict__, objcopy.__dict__, msg=msg)
 | |
|         if hasattr(obj, '__slots__'):
 | |
|             self.assertListEqual(obj.__slots__, objcopy.__slots__, msg=msg)
 | |
|             for slot in obj.__slots__:
 | |
|                 self.assertEqual(
 | |
|                     hasattr(obj, slot), hasattr(objcopy, slot), msg=msg)
 | |
|                 self.assertEqual(getattr(obj, slot, None),
 | |
|                                  getattr(objcopy, slot, None), msg=msg)
 | |
| 
 | |
|     def check_unpickling_error(self, errors, data):
 | |
|         with self.subTest(data=data), \
 | |
|              self.assertRaises(errors):
 | |
|             try:
 | |
|                 self.loads(data)
 | |
|             except BaseException as exc:
 | |
|                 if support.verbose > 1:
 | |
|                     print('%-32r - %s: %s' %
 | |
|                           (data, exc.__class__.__name__, exc))
 | |
|                 raise
 | |
| 
 | |
|     def test_load_from_data0(self):
 | |
|         self.assert_is_copy(self._testdata, self.loads(DATA0))
 | |
| 
 | |
|     def test_load_from_data1(self):
 | |
|         self.assert_is_copy(self._testdata, self.loads(DATA1))
 | |
| 
 | |
|     def test_load_from_data2(self):
 | |
|         self.assert_is_copy(self._testdata, self.loads(DATA2))
 | |
| 
 | |
|     def test_load_from_data3(self):
 | |
|         self.assert_is_copy(self._testdata, self.loads(DATA3))
 | |
| 
 | |
|     def test_load_from_data4(self):
 | |
|         self.assert_is_copy(self._testdata, self.loads(DATA4))
 | |
| 
 | |
|     def test_load_classic_instance(self):
 | |
|         # See issue5180.  Test loading 2.x pickles that
 | |
|         # contain an instance of old style class.
 | |
|         for X, args in [(C, ()), (D, ('x',)), (E, ())]:
 | |
|             xname = X.__name__.encode('ascii')
 | |
|             # Protocol 0 (text mode pickle):
 | |
|             """
 | |
|              0: (    MARK
 | |
|              1: i        INST       '__main__ X' (MARK at 0)
 | |
|             13: p    PUT        0
 | |
|             16: (    MARK
 | |
|             17: d        DICT       (MARK at 16)
 | |
|             18: p    PUT        1
 | |
|             21: b    BUILD
 | |
|             22: .    STOP
 | |
|             """
 | |
|             pickle0 = (b"(i__main__\n"
 | |
|                        b"X\n"
 | |
|                        b"p0\n"
 | |
|                        b"(dp1\nb.").replace(b'X', xname)
 | |
|             self.assert_is_copy(X(*args), self.loads(pickle0))
 | |
| 
 | |
|             # Protocol 1 (binary mode pickle)
 | |
|             """
 | |
|              0: (    MARK
 | |
|              1: c        GLOBAL     '__main__ X'
 | |
|             13: q        BINPUT     0
 | |
|             15: o        OBJ        (MARK at 0)
 | |
|             16: q    BINPUT     1
 | |
|             18: }    EMPTY_DICT
 | |
|             19: q    BINPUT     2
 | |
|             21: b    BUILD
 | |
|             22: .    STOP
 | |
|             """
 | |
|             pickle1 = (b'(c__main__\n'
 | |
|                        b'X\n'
 | |
|                        b'q\x00oq\x01}q\x02b.').replace(b'X', xname)
 | |
|             self.assert_is_copy(X(*args), self.loads(pickle1))
 | |
| 
 | |
|             # Protocol 2 (pickle2 = b'\x80\x02' + pickle1)
 | |
|             """
 | |
|              0: \x80 PROTO      2
 | |
|              2: (    MARK
 | |
|              3: c        GLOBAL     '__main__ X'
 | |
|             15: q        BINPUT     0
 | |
|             17: o        OBJ        (MARK at 2)
 | |
|             18: q    BINPUT     1
 | |
|             20: }    EMPTY_DICT
 | |
|             21: q    BINPUT     2
 | |
|             23: b    BUILD
 | |
|             24: .    STOP
 | |
|             """
 | |
|             pickle2 = (b'\x80\x02(c__main__\n'
 | |
|                        b'X\n'
 | |
|                        b'q\x00oq\x01}q\x02b.').replace(b'X', xname)
 | |
|             self.assert_is_copy(X(*args), self.loads(pickle2))
 | |
| 
 | |
|     def test_maxint64(self):
 | |
|         maxint64 = (1 << 63) - 1
 | |
|         data = b'I' + str(maxint64).encode("ascii") + b'\n.'
 | |
|         got = self.loads(data)
 | |
|         self.assert_is_copy(maxint64, got)
 | |
| 
 | |
|         # Try too with a bogus literal.
 | |
|         data = b'I' + str(maxint64).encode("ascii") + b'JUNK\n.'
 | |
|         self.check_unpickling_error(ValueError, data)
 | |
| 
 | |
|     def test_unpickle_from_2x(self):
 | |
|         # Unpickle non-trivial data from Python 2.x.
 | |
|         loaded = self.loads(DATA_SET)
 | |
|         self.assertEqual(loaded, set([1, 2]))
 | |
|         loaded = self.loads(DATA_XRANGE)
 | |
|         self.assertEqual(type(loaded), type(range(0)))
 | |
|         self.assertEqual(list(loaded), list(range(5)))
 | |
|         loaded = self.loads(DATA_COOKIE)
 | |
|         self.assertEqual(type(loaded), SimpleCookie)
 | |
|         self.assertEqual(list(loaded.keys()), ["key"])
 | |
|         self.assertEqual(loaded["key"].value, "value")
 | |
| 
 | |
|         # Exception objects without arguments pickled from 2.x with protocol 2
 | |
|         for exc in python2_exceptions_without_args:
 | |
|             data = exception_pickle.replace(b'?', exc.__name__.encode("ascii"))
 | |
|             loaded = self.loads(data)
 | |
|             self.assertIs(type(loaded), exc)
 | |
| 
 | |
|         # StandardError is mapped to Exception, test that separately
 | |
|         loaded = self.loads(exception_pickle.replace(b'?', b'StandardError'))
 | |
|         self.assertIs(type(loaded), Exception)
 | |
| 
 | |
|         loaded = self.loads(DATA_UEERR)
 | |
|         self.assertIs(type(loaded), UnicodeEncodeError)
 | |
|         self.assertEqual(loaded.object, "foo")
 | |
|         self.assertEqual(loaded.encoding, "ascii")
 | |
|         self.assertEqual(loaded.start, 0)
 | |
|         self.assertEqual(loaded.end, 1)
 | |
|         self.assertEqual(loaded.reason, "bad")
 | |
| 
 | |
|     def test_load_python2_str_as_bytes(self):
 | |
|         # From Python 2: pickle.dumps('a\x00\xa0', protocol=0)
 | |
|         self.assertEqual(self.loads(b"S'a\\x00\\xa0'\n.",
 | |
|                                     encoding="bytes"), b'a\x00\xa0')
 | |
|         # From Python 2: pickle.dumps('a\x00\xa0', protocol=1)
 | |
|         self.assertEqual(self.loads(b'U\x03a\x00\xa0.',
 | |
|                                     encoding="bytes"), b'a\x00\xa0')
 | |
|         # From Python 2: pickle.dumps('a\x00\xa0', protocol=2)
 | |
|         self.assertEqual(self.loads(b'\x80\x02U\x03a\x00\xa0.',
 | |
|                                     encoding="bytes"), b'a\x00\xa0')
 | |
| 
 | |
|     def test_load_python2_unicode_as_str(self):
 | |
|         # From Python 2: pickle.dumps(u'π', protocol=0)
 | |
|         self.assertEqual(self.loads(b'V\\u03c0\n.',
 | |
|                                     encoding='bytes'), 'π')
 | |
|         # From Python 2: pickle.dumps(u'π', protocol=1)
 | |
|         self.assertEqual(self.loads(b'X\x02\x00\x00\x00\xcf\x80.',
 | |
|                                     encoding="bytes"), 'π')
 | |
|         # From Python 2: pickle.dumps(u'π', protocol=2)
 | |
|         self.assertEqual(self.loads(b'\x80\x02X\x02\x00\x00\x00\xcf\x80.',
 | |
|                                     encoding="bytes"), 'π')
 | |
| 
 | |
|     def test_load_long_python2_str_as_bytes(self):
 | |
|         # From Python 2: pickle.dumps('x' * 300, protocol=1)
 | |
|         self.assertEqual(self.loads(pickle.BINSTRING +
 | |
|                                     struct.pack("<I", 300) +
 | |
|                                     b'x' * 300 + pickle.STOP,
 | |
|                                     encoding='bytes'), b'x' * 300)
 | |
| 
 | |
|     def test_constants(self):
 | |
|         self.assertIsNone(self.loads(b'N.'))
 | |
|         self.assertIs(self.loads(b'\x88.'), True)
 | |
|         self.assertIs(self.loads(b'\x89.'), False)
 | |
|         self.assertIs(self.loads(b'I01\n.'), True)
 | |
|         self.assertIs(self.loads(b'I00\n.'), False)
 | |
| 
 | |
|     def test_empty_bytestring(self):
 | |
|         # issue 11286
 | |
|         empty = self.loads(b'\x80\x03U\x00q\x00.', encoding='koi8-r')
 | |
|         self.assertEqual(empty, '')
 | |
| 
 | |
|     def test_short_binbytes(self):
 | |
|         dumped = b'\x80\x03C\x04\xe2\x82\xac\x00.'
 | |
|         self.assertEqual(self.loads(dumped), b'\xe2\x82\xac\x00')
 | |
| 
 | |
|     def test_binbytes(self):
 | |
|         dumped = b'\x80\x03B\x04\x00\x00\x00\xe2\x82\xac\x00.'
 | |
|         self.assertEqual(self.loads(dumped), b'\xe2\x82\xac\x00')
 | |
| 
 | |
|     @requires_32b
 | |
|     def test_negative_32b_binbytes(self):
 | |
|         # On 32-bit builds, a BINBYTES of 2**31 or more is refused
 | |
|         dumped = b'\x80\x03B\xff\xff\xff\xffxyzq\x00.'
 | |
|         self.check_unpickling_error((pickle.UnpicklingError, OverflowError),
 | |
|                                     dumped)
 | |
| 
 | |
|     @requires_32b
 | |
|     def test_negative_32b_binunicode(self):
 | |
|         # On 32-bit builds, a BINUNICODE of 2**31 or more is refused
 | |
|         dumped = b'\x80\x03X\xff\xff\xff\xffxyzq\x00.'
 | |
|         self.check_unpickling_error((pickle.UnpicklingError, OverflowError),
 | |
|                                     dumped)
 | |
| 
 | |
|     def test_short_binunicode(self):
 | |
|         dumped = b'\x80\x04\x8c\x04\xe2\x82\xac\x00.'
 | |
|         self.assertEqual(self.loads(dumped), '\u20ac\x00')
 | |
| 
 | |
|     def test_misc_get(self):
 | |
|         self.check_unpickling_error(pickle.UnpicklingError, b'g0\np0')
 | |
|         self.check_unpickling_error(pickle.UnpicklingError, b'jens:')
 | |
|         self.check_unpickling_error(pickle.UnpicklingError, b'hens:')
 | |
|         self.assert_is_copy([(100,), (100,)],
 | |
|                             self.loads(b'((Kdtp0\nh\x00l.))'))
 | |
| 
 | |
|     def test_binbytes8(self):
 | |
|         dumped = b'\x80\x04\x8e\4\0\0\0\0\0\0\0\xe2\x82\xac\x00.'
 | |
|         self.assertEqual(self.loads(dumped), b'\xe2\x82\xac\x00')
 | |
| 
 | |
|     def test_binunicode8(self):
 | |
|         dumped = b'\x80\x04\x8d\4\0\0\0\0\0\0\0\xe2\x82\xac\x00.'
 | |
|         self.assertEqual(self.loads(dumped), '\u20ac\x00')
 | |
| 
 | |
|     def test_bytearray8(self):
 | |
|         dumped = b'\x80\x05\x96\x03\x00\x00\x00\x00\x00\x00\x00xxx.'
 | |
|         self.assertEqual(self.loads(dumped), bytearray(b'xxx'))
 | |
| 
 | |
|     @requires_32b
 | |
|     def test_large_32b_binbytes8(self):
 | |
|         dumped = b'\x80\x04\x8e\4\0\0\0\1\0\0\0\xe2\x82\xac\x00.'
 | |
|         self.check_unpickling_error((pickle.UnpicklingError, OverflowError),
 | |
|                                     dumped)
 | |
| 
 | |
|     @requires_32b
 | |
|     def test_large_32b_bytearray8(self):
 | |
|         dumped = b'\x80\x05\x96\4\0\0\0\1\0\0\0\xe2\x82\xac\x00.'
 | |
|         self.check_unpickling_error((pickle.UnpicklingError, OverflowError),
 | |
|                                     dumped)
 | |
| 
 | |
|     @requires_32b
 | |
|     def test_large_32b_binunicode8(self):
 | |
|         dumped = b'\x80\x04\x8d\4\0\0\0\1\0\0\0\xe2\x82\xac\x00.'
 | |
|         self.check_unpickling_error((pickle.UnpicklingError, OverflowError),
 | |
|                                     dumped)
 | |
| 
 | |
|     def test_get(self):
 | |
|         pickled = b'((lp100000\ng100000\nt.'
 | |
|         unpickled = self.loads(pickled)
 | |
|         self.assertEqual(unpickled, ([],)*2)
 | |
|         self.assertIs(unpickled[0], unpickled[1])
 | |
| 
 | |
|     def test_binget(self):
 | |
|         pickled = b'(]q\xffh\xfft.'
 | |
|         unpickled = self.loads(pickled)
 | |
|         self.assertEqual(unpickled, ([],)*2)
 | |
|         self.assertIs(unpickled[0], unpickled[1])
 | |
| 
 | |
|     def test_long_binget(self):
 | |
|         pickled = b'(]r\x00\x00\x01\x00j\x00\x00\x01\x00t.'
 | |
|         unpickled = self.loads(pickled)
 | |
|         self.assertEqual(unpickled, ([],)*2)
 | |
|         self.assertIs(unpickled[0], unpickled[1])
 | |
| 
 | |
|     def test_dup(self):
 | |
|         pickled = b'((l2t.'
 | |
|         unpickled = self.loads(pickled)
 | |
|         self.assertEqual(unpickled, ([],)*2)
 | |
|         self.assertIs(unpickled[0], unpickled[1])
 | |
| 
 | |
|     def test_negative_put(self):
 | |
|         # Issue #12847
 | |
|         dumped = b'Va\np-1\n.'
 | |
|         self.check_unpickling_error(ValueError, dumped)
 | |
| 
 | |
|     @requires_32b
 | |
|     def test_negative_32b_binput(self):
 | |
|         # Issue #12847
 | |
|         dumped = b'\x80\x03X\x01\x00\x00\x00ar\xff\xff\xff\xff.'
 | |
|         self.check_unpickling_error(ValueError, dumped)
 | |
| 
 | |
|     def test_badly_escaped_string(self):
 | |
|         self.check_unpickling_error(ValueError, b"S'\\'\n.")
 | |
| 
 | |
|     def test_badly_quoted_string(self):
 | |
|         # Issue #17710
 | |
|         badpickles = [b"S'\n.",
 | |
|                       b'S"\n.',
 | |
|                       b'S\' \n.',
 | |
|                       b'S" \n.',
 | |
|                       b'S\'"\n.',
 | |
|                       b'S"\'\n.',
 | |
|                       b"S' ' \n.",
 | |
|                       b'S" " \n.',
 | |
|                       b"S ''\n.",
 | |
|                       b'S ""\n.',
 | |
|                       b'S \n.',
 | |
|                       b'S\n.',
 | |
|                       b'S.']
 | |
|         for p in badpickles:
 | |
|             self.check_unpickling_error(pickle.UnpicklingError, p)
 | |
| 
 | |
|     def test_correctly_quoted_string(self):
 | |
|         goodpickles = [(b"S''\n.", ''),
 | |
|                        (b'S""\n.', ''),
 | |
|                        (b'S"\\n"\n.', '\n'),
 | |
|                        (b"S'\\n'\n.", '\n')]
 | |
|         for p, expected in goodpickles:
 | |
|             self.assertEqual(self.loads(p), expected)
 | |
| 
 | |
|     def test_frame_readline(self):
 | |
|         pickled = b'\x80\x04\x95\x05\x00\x00\x00\x00\x00\x00\x00I42\n.'
 | |
|         #    0: \x80 PROTO      4
 | |
|         #    2: \x95 FRAME      5
 | |
|         #   11: I    INT        42
 | |
|         #   15: .    STOP
 | |
|         self.assertEqual(self.loads(pickled), 42)
 | |
| 
 | |
|     def test_compat_unpickle(self):
 | |
|         # xrange(1, 7)
 | |
|         pickled = b'\x80\x02c__builtin__\nxrange\nK\x01K\x07K\x01\x87R.'
 | |
|         unpickled = self.loads(pickled)
 | |
|         self.assertIs(type(unpickled), range)
 | |
|         self.assertEqual(unpickled, range(1, 7))
 | |
|         self.assertEqual(list(unpickled), [1, 2, 3, 4, 5, 6])
 | |
|         # reduce
 | |
|         pickled = b'\x80\x02c__builtin__\nreduce\n.'
 | |
|         self.assertIs(self.loads(pickled), functools.reduce)
 | |
|         # whichdb.whichdb
 | |
|         pickled = b'\x80\x02cwhichdb\nwhichdb\n.'
 | |
|         self.assertIs(self.loads(pickled), dbm.whichdb)
 | |
|         # Exception(), StandardError()
 | |
|         for name in (b'Exception', b'StandardError'):
 | |
|             pickled = (b'\x80\x02cexceptions\n' + name + b'\nU\x03ugh\x85R.')
 | |
|             unpickled = self.loads(pickled)
 | |
|             self.assertIs(type(unpickled), Exception)
 | |
|             self.assertEqual(str(unpickled), 'ugh')
 | |
|         # UserDict.UserDict({1: 2}), UserDict.IterableUserDict({1: 2})
 | |
|         for name in (b'UserDict', b'IterableUserDict'):
 | |
|             pickled = (b'\x80\x02(cUserDict\n' + name +
 | |
|                        b'\no}U\x04data}K\x01K\x02ssb.')
 | |
|             unpickled = self.loads(pickled)
 | |
|             self.assertIs(type(unpickled), collections.UserDict)
 | |
|             self.assertEqual(unpickled, collections.UserDict({1: 2}))
 | |
| 
 | |
|     def test_bad_reduce(self):
 | |
|         self.assertEqual(self.loads(b'cbuiltins\nint\n)R.'), 0)
 | |
|         self.check_unpickling_error(TypeError, b'N)R.')
 | |
|         self.check_unpickling_error(TypeError, b'cbuiltins\nint\nNR.')
 | |
| 
 | |
|     def test_bad_newobj(self):
 | |
|         error = (pickle.UnpicklingError, TypeError)
 | |
|         self.assertEqual(self.loads(b'cbuiltins\nint\n)\x81.'), 0)
 | |
|         self.check_unpickling_error(error, b'cbuiltins\nlen\n)\x81.')
 | |
|         self.check_unpickling_error(error, b'cbuiltins\nint\nN\x81.')
 | |
| 
 | |
|     def test_bad_newobj_ex(self):
 | |
|         error = (pickle.UnpicklingError, TypeError)
 | |
|         self.assertEqual(self.loads(b'cbuiltins\nint\n)}\x92.'), 0)
 | |
|         self.check_unpickling_error(error, b'cbuiltins\nlen\n)}\x92.')
 | |
|         self.check_unpickling_error(error, b'cbuiltins\nint\nN}\x92.')
 | |
|         self.check_unpickling_error(error, b'cbuiltins\nint\n)N\x92.')
 | |
| 
 | |
|     def test_bad_stack(self):
 | |
|         badpickles = [
 | |
|             b'.',                       # STOP
 | |
|             b'0',                       # POP
 | |
|             b'1',                       # POP_MARK
 | |
|             b'2',                       # DUP
 | |
|             b'(2',
 | |
|             b'R',                       # REDUCE
 | |
|             b')R',
 | |
|             b'a',                       # APPEND
 | |
|             b'Na',
 | |
|             b'b',                       # BUILD
 | |
|             b'Nb',
 | |
|             b'd',                       # DICT
 | |
|             b'e',                       # APPENDS
 | |
|             b'(e',
 | |
|             b'ibuiltins\nlist\n',       # INST
 | |
|             b'l',                       # LIST
 | |
|             b'o',                       # OBJ
 | |
|             b'(o',
 | |
|             b'p1\n',                    # PUT
 | |
|             b'q\x00',                   # BINPUT
 | |
|             b'r\x00\x00\x00\x00',       # LONG_BINPUT
 | |
|             b's',                       # SETITEM
 | |
|             b'Ns',
 | |
|             b'NNs',
 | |
|             b't',                       # TUPLE
 | |
|             b'u',                       # SETITEMS
 | |
|             b'(u',
 | |
|             b'}(Nu',
 | |
|             b'\x81',                    # NEWOBJ
 | |
|             b')\x81',
 | |
|             b'\x85',                    # TUPLE1
 | |
|             b'\x86',                    # TUPLE2
 | |
|             b'N\x86',
 | |
|             b'\x87',                    # TUPLE3
 | |
|             b'N\x87',
 | |
|             b'NN\x87',
 | |
|             b'\x90',                    # ADDITEMS
 | |
|             b'(\x90',
 | |
|             b'\x91',                    # FROZENSET
 | |
|             b'\x92',                    # NEWOBJ_EX
 | |
|             b')}\x92',
 | |
|             b'\x93',                    # STACK_GLOBAL
 | |
|             b'Vlist\n\x93',
 | |
|             b'\x94',                    # MEMOIZE
 | |
|         ]
 | |
|         for p in badpickles:
 | |
|             self.check_unpickling_error(self.bad_stack_errors, p)
 | |
| 
 | |
|     def test_bad_mark(self):
 | |
|         badpickles = [
 | |
|             b'N(.',                     # STOP
 | |
|             b'N(2',                     # DUP
 | |
|             b'cbuiltins\nlist\n)(R',    # REDUCE
 | |
|             b'cbuiltins\nlist\n()R',
 | |
|             b']N(a',                    # APPEND
 | |
|                                         # BUILD
 | |
|             b'cbuiltins\nValueError\n)R}(b',
 | |
|             b'cbuiltins\nValueError\n)R(}b',
 | |
|             b'(Nd',                     # DICT
 | |
|             b'N(p1\n',                  # PUT
 | |
|             b'N(q\x00',                 # BINPUT
 | |
|             b'N(r\x00\x00\x00\x00',     # LONG_BINPUT
 | |
|             b'}NN(s',                   # SETITEM
 | |
|             b'}N(Ns',
 | |
|             b'}(NNs',
 | |
|             b'}((u',                    # SETITEMS
 | |
|             b'cbuiltins\nlist\n)(\x81', # NEWOBJ
 | |
|             b'cbuiltins\nlist\n()\x81',
 | |
|             b'N(\x85',                  # TUPLE1
 | |
|             b'NN(\x86',                 # TUPLE2
 | |
|             b'N(N\x86',
 | |
|             b'NNN(\x87',                # TUPLE3
 | |
|             b'NN(N\x87',
 | |
|             b'N(NN\x87',
 | |
|             b']((\x90',                 # ADDITEMS
 | |
|                                         # NEWOBJ_EX
 | |
|             b'cbuiltins\nlist\n)}(\x92',
 | |
|             b'cbuiltins\nlist\n)(}\x92',
 | |
|             b'cbuiltins\nlist\n()}\x92',
 | |
|                                         # STACK_GLOBAL
 | |
|             b'Vbuiltins\n(Vlist\n\x93',
 | |
|             b'Vbuiltins\nVlist\n(\x93',
 | |
|             b'N(\x94',                  # MEMOIZE
 | |
|         ]
 | |
|         for p in badpickles:
 | |
|             self.check_unpickling_error(self.bad_stack_errors, p)
 | |
| 
 | |
|     def test_truncated_data(self):
 | |
|         self.check_unpickling_error(EOFError, b'')
 | |
|         self.check_unpickling_error(EOFError, b'N')
 | |
|         badpickles = [
 | |
|             b'B',                       # BINBYTES
 | |
|             b'B\x03\x00\x00',
 | |
|             b'B\x03\x00\x00\x00',
 | |
|             b'B\x03\x00\x00\x00ab',
 | |
|             b'C',                       # SHORT_BINBYTES
 | |
|             b'C\x03',
 | |
|             b'C\x03ab',
 | |
|             b'F',                       # FLOAT
 | |
|             b'F0.0',
 | |
|             b'F0.00',
 | |
|             b'G',                       # BINFLOAT
 | |
|             b'G\x00\x00\x00\x00\x00\x00\x00',
 | |
|             b'I',                       # INT
 | |
|             b'I0',
 | |
|             b'J',                       # BININT
 | |
|             b'J\x00\x00\x00',
 | |
|             b'K',                       # BININT1
 | |
|             b'L',                       # LONG
 | |
|             b'L0',
 | |
|             b'L10',
 | |
|             b'L0L',
 | |
|             b'L10L',
 | |
|             b'M',                       # BININT2
 | |
|             b'M\x00',
 | |
|             # b'P',                       # PERSID
 | |
|             # b'Pabc',
 | |
|             b'S',                       # STRING
 | |
|             b"S'abc'",
 | |
|             b'T',                       # BINSTRING
 | |
|             b'T\x03\x00\x00',
 | |
|             b'T\x03\x00\x00\x00',
 | |
|             b'T\x03\x00\x00\x00ab',
 | |
|             b'U',                       # SHORT_BINSTRING
 | |
|             b'U\x03',
 | |
|             b'U\x03ab',
 | |
|             b'V',                       # UNICODE
 | |
|             b'Vabc',
 | |
|             b'X',                       # BINUNICODE
 | |
|             b'X\x03\x00\x00',
 | |
|             b'X\x03\x00\x00\x00',
 | |
|             b'X\x03\x00\x00\x00ab',
 | |
|             b'(c',                      # GLOBAL
 | |
|             b'(cbuiltins',
 | |
|             b'(cbuiltins\n',
 | |
|             b'(cbuiltins\nlist',
 | |
|             b'Ng',                      # GET
 | |
|             b'Ng0',
 | |
|             b'(i',                      # INST
 | |
|             b'(ibuiltins',
 | |
|             b'(ibuiltins\n',
 | |
|             b'(ibuiltins\nlist',
 | |
|             b'Nh',                      # BINGET
 | |
|             b'Nj',                      # LONG_BINGET
 | |
|             b'Nj\x00\x00\x00',
 | |
|             b'Np',                      # PUT
 | |
|             b'Np0',
 | |
|             b'Nq',                      # BINPUT
 | |
|             b'Nr',                      # LONG_BINPUT
 | |
|             b'Nr\x00\x00\x00',
 | |
|             b'\x80',                    # PROTO
 | |
|             b'\x82',                    # EXT1
 | |
|             b'\x83',                    # EXT2
 | |
|             b'\x84\x01',
 | |
|             b'\x84',                    # EXT4
 | |
|             b'\x84\x01\x00\x00',
 | |
|             b'\x8a',                    # LONG1
 | |
|             b'\x8b',                    # LONG4
 | |
|             b'\x8b\x00\x00\x00',
 | |
|             b'\x8c',                    # SHORT_BINUNICODE
 | |
|             b'\x8c\x03',
 | |
|             b'\x8c\x03ab',
 | |
|             b'\x8d',                    # BINUNICODE8
 | |
|             b'\x8d\x03\x00\x00\x00\x00\x00\x00',
 | |
|             b'\x8d\x03\x00\x00\x00\x00\x00\x00\x00',
 | |
|             b'\x8d\x03\x00\x00\x00\x00\x00\x00\x00ab',
 | |
|             b'\x8e',                    # BINBYTES8
 | |
|             b'\x8e\x03\x00\x00\x00\x00\x00\x00',
 | |
|             b'\x8e\x03\x00\x00\x00\x00\x00\x00\x00',
 | |
|             b'\x8e\x03\x00\x00\x00\x00\x00\x00\x00ab',
 | |
|             b'\x96',                    # BYTEARRAY8
 | |
|             b'\x96\x03\x00\x00\x00\x00\x00\x00',
 | |
|             b'\x96\x03\x00\x00\x00\x00\x00\x00\x00',
 | |
|             b'\x96\x03\x00\x00\x00\x00\x00\x00\x00ab',
 | |
|             b'\x95',                    # FRAME
 | |
|             b'\x95\x02\x00\x00\x00\x00\x00\x00',
 | |
|             b'\x95\x02\x00\x00\x00\x00\x00\x00\x00',
 | |
|             b'\x95\x02\x00\x00\x00\x00\x00\x00\x00N',
 | |
|         ]
 | |
|         for p in badpickles:
 | |
|             self.check_unpickling_error(self.truncated_errors, p)
 | |
| 
 | |
|     @reap_threads
 | |
|     def test_unpickle_module_race(self):
 | |
|         # https://bugs.python.org/issue34572
 | |
|         locker_module = dedent("""
 | |
|         import threading
 | |
|         barrier = threading.Barrier(2)
 | |
|         """)
 | |
|         locking_import_module = dedent("""
 | |
|         import locker
 | |
|         locker.barrier.wait()
 | |
|         class ToBeUnpickled(object):
 | |
|             pass
 | |
|         """)
 | |
| 
 | |
|         os.mkdir(TESTFN)
 | |
|         self.addCleanup(shutil.rmtree, TESTFN)
 | |
|         sys.path.insert(0, TESTFN)
 | |
|         self.addCleanup(sys.path.remove, TESTFN)
 | |
|         with open(os.path.join(TESTFN, "locker.py"), "wb") as f:
 | |
|             f.write(locker_module.encode('utf-8'))
 | |
|         with open(os.path.join(TESTFN, "locking_import.py"), "wb") as f:
 | |
|             f.write(locking_import_module.encode('utf-8'))
 | |
|         self.addCleanup(forget, "locker")
 | |
|         self.addCleanup(forget, "locking_import")
 | |
| 
 | |
|         import locker
 | |
| 
 | |
|         pickle_bytes = (
 | |
|             b'\x80\x03clocking_import\nToBeUnpickled\nq\x00)\x81q\x01.')
 | |
| 
 | |
|         # Then try to unpickle two of these simultaneously
 | |
|         # One of them will cause the module import, and we want it to block
 | |
|         # until the other one either:
 | |
|         #   - fails (before the patch for this issue)
 | |
|         #   - blocks on the import lock for the module, as it should
 | |
|         results = []
 | |
|         barrier = threading.Barrier(3)
 | |
|         def t():
 | |
|             # This ensures the threads have all started
 | |
|             # presumably barrier release is faster than thread startup
 | |
|             barrier.wait()
 | |
|             results.append(pickle.loads(pickle_bytes))
 | |
| 
 | |
|         t1 = threading.Thread(target=t)
 | |
|         t2 = threading.Thread(target=t)
 | |
|         t1.start()
 | |
|         t2.start()
 | |
| 
 | |
|         barrier.wait()
 | |
|         # could have delay here
 | |
|         locker.barrier.wait()
 | |
| 
 | |
|         t1.join()
 | |
|         t2.join()
 | |
| 
 | |
|         from locking_import import ToBeUnpickled
 | |
|         self.assertEqual(
 | |
|             [type(x) for x in results],
 | |
|             [ToBeUnpickled] * 2)
 | |
| 
 | |
| 
 | |
| 
 | |
| class AbstractPickleTests(unittest.TestCase):
 | |
|     # Subclass must define self.dumps, self.loads.
 | |
| 
 | |
|     optimized = False
 | |
| 
 | |
|     _testdata = AbstractUnpickleTests._testdata
 | |
| 
 | |
|     def setUp(self):
 | |
|         pass
 | |
| 
 | |
|     assert_is_copy = AbstractUnpickleTests.assert_is_copy
 | |
| 
 | |
|     def test_misc(self):
 | |
|         # test various datatypes not tested by testdata
 | |
|         for proto in protocols:
 | |
|             x = myint(4)
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assert_is_copy(x, y)
 | |
| 
 | |
|             x = (1, ())
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assert_is_copy(x, y)
 | |
| 
 | |
|             x = initarg(1, x)
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assert_is_copy(x, y)
 | |
| 
 | |
|         # XXX test __reduce__ protocol?
 | |
| 
 | |
|     def test_roundtrip_equality(self):
 | |
|         expected = self._testdata
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(expected, proto)
 | |
|             got = self.loads(s)
 | |
|             self.assert_is_copy(expected, got)
 | |
| 
 | |
|     # There are gratuitous differences between pickles produced by
 | |
|     # pickle and cPickle, largely because cPickle starts PUT indices at
 | |
|     # 1 and pickle starts them at 0.  See XXX comment in cPickle's put2() --
 | |
|     # there's a comment with an exclamation point there whose meaning
 | |
|     # is a mystery.  cPickle also suppresses PUT for objects with a refcount
 | |
|     # of 1.
 | |
|     def dont_test_disassembly(self):
 | |
|         from io import StringIO
 | |
|         from pickletools import dis
 | |
| 
 | |
|         for proto, expected in (0, DATA0_DIS), (1, DATA1_DIS):
 | |
|             s = self.dumps(self._testdata, proto)
 | |
|             filelike = StringIO()
 | |
|             dis(s, out=filelike)
 | |
|             got = filelike.getvalue()
 | |
|             self.assertEqual(expected, got)
 | |
| 
 | |
|     def _test_recursive_list(self, cls, aslist=identity, minprotocol=0):
 | |
|         # List containing itself.
 | |
|         l = cls()
 | |
|         l.append(l)
 | |
|         for proto in range(minprotocol, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             s = self.dumps(l, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, cls)
 | |
|             y = aslist(x)
 | |
|             self.assertEqual(len(y), 1)
 | |
|             self.assertIs(y[0], x)
 | |
| 
 | |
|     def test_recursive_list(self):
 | |
|         self._test_recursive_list(list)
 | |
| 
 | |
|     def test_recursive_list_subclass(self):
 | |
|         self._test_recursive_list(MyList, minprotocol=2)
 | |
| 
 | |
|     def test_recursive_list_like(self):
 | |
|         self._test_recursive_list(REX_six, aslist=lambda x: x.items)
 | |
| 
 | |
|     def _test_recursive_tuple_and_list(self, cls, aslist=identity, minprotocol=0):
 | |
|         # Tuple containing a list containing the original tuple.
 | |
|         t = (cls(),)
 | |
|         t[0].append(t)
 | |
|         for proto in range(minprotocol, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             s = self.dumps(t, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, tuple)
 | |
|             self.assertEqual(len(x), 1)
 | |
|             self.assertIsInstance(x[0], cls)
 | |
|             y = aslist(x[0])
 | |
|             self.assertEqual(len(y), 1)
 | |
|             self.assertIs(y[0], x)
 | |
| 
 | |
|         # List containing a tuple containing the original list.
 | |
|         t, = t
 | |
|         for proto in range(minprotocol, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             s = self.dumps(t, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, cls)
 | |
|             y = aslist(x)
 | |
|             self.assertEqual(len(y), 1)
 | |
|             self.assertIsInstance(y[0], tuple)
 | |
|             self.assertEqual(len(y[0]), 1)
 | |
|             self.assertIs(y[0][0], x)
 | |
| 
 | |
|     def test_recursive_tuple_and_list(self):
 | |
|         self._test_recursive_tuple_and_list(list)
 | |
| 
 | |
|     def test_recursive_tuple_and_list_subclass(self):
 | |
|         self._test_recursive_tuple_and_list(MyList, minprotocol=2)
 | |
| 
 | |
|     def test_recursive_tuple_and_list_like(self):
 | |
|         self._test_recursive_tuple_and_list(REX_six, aslist=lambda x: x.items)
 | |
| 
 | |
|     def _test_recursive_dict(self, cls, asdict=identity, minprotocol=0):
 | |
|         # Dict containing itself.
 | |
|         d = cls()
 | |
|         d[1] = d
 | |
|         for proto in range(minprotocol, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             s = self.dumps(d, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, cls)
 | |
|             y = asdict(x)
 | |
|             self.assertEqual(list(y.keys()), [1])
 | |
|             self.assertIs(y[1], x)
 | |
| 
 | |
|     def test_recursive_dict(self):
 | |
|         self._test_recursive_dict(dict)
 | |
| 
 | |
|     def test_recursive_dict_subclass(self):
 | |
|         self._test_recursive_dict(MyDict, minprotocol=2)
 | |
| 
 | |
|     def test_recursive_dict_like(self):
 | |
|         self._test_recursive_dict(REX_seven, asdict=lambda x: x.table)
 | |
| 
 | |
|     def _test_recursive_tuple_and_dict(self, cls, asdict=identity, minprotocol=0):
 | |
|         # Tuple containing a dict containing the original tuple.
 | |
|         t = (cls(),)
 | |
|         t[0][1] = t
 | |
|         for proto in range(minprotocol, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             s = self.dumps(t, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, tuple)
 | |
|             self.assertEqual(len(x), 1)
 | |
|             self.assertIsInstance(x[0], cls)
 | |
|             y = asdict(x[0])
 | |
|             self.assertEqual(list(y), [1])
 | |
|             self.assertIs(y[1], x)
 | |
| 
 | |
|         # Dict containing a tuple containing the original dict.
 | |
|         t, = t
 | |
|         for proto in range(minprotocol, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             s = self.dumps(t, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, cls)
 | |
|             y = asdict(x)
 | |
|             self.assertEqual(list(y), [1])
 | |
|             self.assertIsInstance(y[1], tuple)
 | |
|             self.assertEqual(len(y[1]), 1)
 | |
|             self.assertIs(y[1][0], x)
 | |
| 
 | |
|     def test_recursive_tuple_and_dict(self):
 | |
|         self._test_recursive_tuple_and_dict(dict)
 | |
| 
 | |
|     def test_recursive_tuple_and_dict_subclass(self):
 | |
|         self._test_recursive_tuple_and_dict(MyDict, minprotocol=2)
 | |
| 
 | |
|     def test_recursive_tuple_and_dict_like(self):
 | |
|         self._test_recursive_tuple_and_dict(REX_seven, asdict=lambda x: x.table)
 | |
| 
 | |
|     def _test_recursive_dict_key(self, cls, asdict=identity, minprotocol=0):
 | |
|         # Dict containing an immutable object (as key) containing the original
 | |
|         # dict.
 | |
|         d = cls()
 | |
|         d[K(d)] = 1
 | |
|         for proto in range(minprotocol, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             s = self.dumps(d, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, cls)
 | |
|             y = asdict(x)
 | |
|             self.assertEqual(len(y.keys()), 1)
 | |
|             self.assertIsInstance(list(y.keys())[0], K)
 | |
|             self.assertIs(list(y.keys())[0].value, x)
 | |
| 
 | |
|     def test_recursive_dict_key(self):
 | |
|         self._test_recursive_dict_key(dict)
 | |
| 
 | |
|     def test_recursive_dict_subclass_key(self):
 | |
|         self._test_recursive_dict_key(MyDict, minprotocol=2)
 | |
| 
 | |
|     def test_recursive_dict_like_key(self):
 | |
|         self._test_recursive_dict_key(REX_seven, asdict=lambda x: x.table)
 | |
| 
 | |
|     def _test_recursive_tuple_and_dict_key(self, cls, asdict=identity, minprotocol=0):
 | |
|         # Tuple containing a dict containing an immutable object (as key)
 | |
|         # containing the original tuple.
 | |
|         t = (cls(),)
 | |
|         t[0][K(t)] = 1
 | |
|         for proto in range(minprotocol, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             s = self.dumps(t, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, tuple)
 | |
|             self.assertEqual(len(x), 1)
 | |
|             self.assertIsInstance(x[0], cls)
 | |
|             y = asdict(x[0])
 | |
|             self.assertEqual(len(y), 1)
 | |
|             self.assertIsInstance(list(y.keys())[0], K)
 | |
|             self.assertIs(list(y.keys())[0].value, x)
 | |
| 
 | |
|         # Dict containing an immutable object (as key) containing a tuple
 | |
|         # containing the original dict.
 | |
|         t, = t
 | |
|         for proto in range(minprotocol, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             s = self.dumps(t, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, cls)
 | |
|             y = asdict(x)
 | |
|             self.assertEqual(len(y), 1)
 | |
|             self.assertIsInstance(list(y.keys())[0], K)
 | |
|             self.assertIs(list(y.keys())[0].value[0], x)
 | |
| 
 | |
|     def test_recursive_tuple_and_dict_key(self):
 | |
|         self._test_recursive_tuple_and_dict_key(dict)
 | |
| 
 | |
|     def test_recursive_tuple_and_dict_subclass_key(self):
 | |
|         self._test_recursive_tuple_and_dict_key(MyDict, minprotocol=2)
 | |
| 
 | |
|     def test_recursive_tuple_and_dict_like_key(self):
 | |
|         self._test_recursive_tuple_and_dict_key(REX_seven, asdict=lambda x: x.table)
 | |
| 
 | |
|     def test_recursive_set(self):
 | |
|         # Set containing an immutable object containing the original set.
 | |
|         y = set()
 | |
|         y.add(K(y))
 | |
|         for proto in range(4, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             s = self.dumps(y, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, set)
 | |
|             self.assertEqual(len(x), 1)
 | |
|             self.assertIsInstance(list(x)[0], K)
 | |
|             self.assertIs(list(x)[0].value, x)
 | |
| 
 | |
|         # Immutable object containing a set containing the original object.
 | |
|         y, = y
 | |
|         for proto in range(4, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             s = self.dumps(y, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, K)
 | |
|             self.assertIsInstance(x.value, set)
 | |
|             self.assertEqual(len(x.value), 1)
 | |
|             self.assertIs(list(x.value)[0], x)
 | |
| 
 | |
|     def test_recursive_inst(self):
 | |
|         # Mutable object containing itself.
 | |
|         i = Object()
 | |
|         i.attr = i
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(i, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, Object)
 | |
|             self.assertEqual(dir(x), dir(i))
 | |
|             self.assertIs(x.attr, x)
 | |
| 
 | |
|     def test_recursive_multi(self):
 | |
|         l = []
 | |
|         d = {1:l}
 | |
|         i = Object()
 | |
|         i.attr = d
 | |
|         l.append(i)
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(l, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, list)
 | |
|             self.assertEqual(len(x), 1)
 | |
|             self.assertEqual(dir(x[0]), dir(i))
 | |
|             self.assertEqual(list(x[0].attr.keys()), [1])
 | |
|             self.assertIs(x[0].attr[1], x)
 | |
| 
 | |
|     def _test_recursive_collection_and_inst(self, factory):
 | |
|         # Mutable object containing a collection containing the original
 | |
|         # object.
 | |
|         o = Object()
 | |
|         o.attr = factory([o])
 | |
|         t = type(o.attr)
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(o, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x.attr, t)
 | |
|             self.assertEqual(len(x.attr), 1)
 | |
|             self.assertIsInstance(list(x.attr)[0], Object)
 | |
|             self.assertIs(list(x.attr)[0], x)
 | |
| 
 | |
|         # Collection containing a mutable object containing the original
 | |
|         # collection.
 | |
|         o = o.attr
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(o, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, t)
 | |
|             self.assertEqual(len(x), 1)
 | |
|             self.assertIsInstance(list(x)[0], Object)
 | |
|             self.assertIs(list(x)[0].attr, x)
 | |
| 
 | |
|     def test_recursive_list_and_inst(self):
 | |
|         self._test_recursive_collection_and_inst(list)
 | |
| 
 | |
|     def test_recursive_tuple_and_inst(self):
 | |
|         self._test_recursive_collection_and_inst(tuple)
 | |
| 
 | |
|     def test_recursive_dict_and_inst(self):
 | |
|         self._test_recursive_collection_and_inst(dict.fromkeys)
 | |
| 
 | |
|     def test_recursive_set_and_inst(self):
 | |
|         self._test_recursive_collection_and_inst(set)
 | |
| 
 | |
|     def test_recursive_frozenset_and_inst(self):
 | |
|         self._test_recursive_collection_and_inst(frozenset)
 | |
| 
 | |
|     def test_recursive_list_subclass_and_inst(self):
 | |
|         self._test_recursive_collection_and_inst(MyList)
 | |
| 
 | |
|     def test_recursive_tuple_subclass_and_inst(self):
 | |
|         self._test_recursive_collection_and_inst(MyTuple)
 | |
| 
 | |
|     def test_recursive_dict_subclass_and_inst(self):
 | |
|         self._test_recursive_collection_and_inst(MyDict.fromkeys)
 | |
| 
 | |
|     def test_recursive_set_subclass_and_inst(self):
 | |
|         self._test_recursive_collection_and_inst(MySet)
 | |
| 
 | |
|     def test_recursive_frozenset_subclass_and_inst(self):
 | |
|         self._test_recursive_collection_and_inst(MyFrozenSet)
 | |
| 
 | |
|     def test_recursive_inst_state(self):
 | |
|         # Mutable object containing itself.
 | |
|         y = REX_state()
 | |
|         y.state = y
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(y, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, REX_state)
 | |
|             self.assertIs(x.state, x)
 | |
| 
 | |
|     def test_recursive_tuple_and_inst_state(self):
 | |
|         # Tuple containing a mutable object containing the original tuple.
 | |
|         t = (REX_state(),)
 | |
|         t[0].state = t
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(t, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, tuple)
 | |
|             self.assertEqual(len(x), 1)
 | |
|             self.assertIsInstance(x[0], REX_state)
 | |
|             self.assertIs(x[0].state, x)
 | |
| 
 | |
|         # Mutable object containing a tuple containing the object.
 | |
|         t, = t
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(t, proto)
 | |
|             x = self.loads(s)
 | |
|             self.assertIsInstance(x, REX_state)
 | |
|             self.assertIsInstance(x.state, tuple)
 | |
|             self.assertEqual(len(x.state), 1)
 | |
|             self.assertIs(x.state[0], x)
 | |
| 
 | |
|     def test_unicode(self):
 | |
|         endcases = ['', '<\\u>', '<\\\u1234>', '<\n>',
 | |
|                     '<\\>', '<\\\U00012345>',
 | |
|                     # surrogates
 | |
|                     '<\udc80>']
 | |
|         for proto in protocols:
 | |
|             for u in endcases:
 | |
|                 p = self.dumps(u, proto)
 | |
|                 u2 = self.loads(p)
 | |
|                 self.assert_is_copy(u, u2)
 | |
| 
 | |
|     def test_unicode_high_plane(self):
 | |
|         t = '\U00012345'
 | |
|         for proto in protocols:
 | |
|             p = self.dumps(t, proto)
 | |
|             t2 = self.loads(p)
 | |
|             self.assert_is_copy(t, t2)
 | |
| 
 | |
|     def test_bytes(self):
 | |
|         for proto in protocols:
 | |
|             for s in b'', b'xyz', b'xyz'*100:
 | |
|                 p = self.dumps(s, proto)
 | |
|                 self.assert_is_copy(s, self.loads(p))
 | |
|             for s in [bytes([i]) for i in range(256)]:
 | |
|                 p = self.dumps(s, proto)
 | |
|                 self.assert_is_copy(s, self.loads(p))
 | |
|             for s in [bytes([i, i]) for i in range(256)]:
 | |
|                 p = self.dumps(s, proto)
 | |
|                 self.assert_is_copy(s, self.loads(p))
 | |
| 
 | |
|     def test_bytearray(self):
 | |
|         for proto in protocols:
 | |
|             for s in b'', b'xyz', b'xyz'*100:
 | |
|                 b = bytearray(s)
 | |
|                 p = self.dumps(b, proto)
 | |
|                 bb = self.loads(p)
 | |
|                 self.assertIsNot(bb, b)
 | |
|                 self.assert_is_copy(b, bb)
 | |
|                 if proto <= 3:
 | |
|                     # bytearray is serialized using a global reference
 | |
|                     self.assertIn(b'bytearray', p)
 | |
|                     self.assertTrue(opcode_in_pickle(pickle.GLOBAL, p))
 | |
|                 elif proto == 4:
 | |
|                     self.assertIn(b'bytearray', p)
 | |
|                     self.assertTrue(opcode_in_pickle(pickle.STACK_GLOBAL, p))
 | |
|                 elif proto == 5:
 | |
|                     self.assertNotIn(b'bytearray', p)
 | |
|                     self.assertTrue(opcode_in_pickle(pickle.BYTEARRAY8, p))
 | |
| 
 | |
|     def test_ints(self):
 | |
|         for proto in protocols:
 | |
|             n = sys.maxsize
 | |
|             while n:
 | |
|                 for expected in (-n, n):
 | |
|                     s = self.dumps(expected, proto)
 | |
|                     n2 = self.loads(s)
 | |
|                     self.assert_is_copy(expected, n2)
 | |
|                 n = n >> 1
 | |
| 
 | |
|     def test_long(self):
 | |
|         for proto in protocols:
 | |
|             # 256 bytes is where LONG4 begins.
 | |
|             for nbits in 1, 8, 8*254, 8*255, 8*256, 8*257:
 | |
|                 nbase = 1 << nbits
 | |
|                 for npos in nbase-1, nbase, nbase+1:
 | |
|                     for n in npos, -npos:
 | |
|                         pickle = self.dumps(n, proto)
 | |
|                         got = self.loads(pickle)
 | |
|                         self.assert_is_copy(n, got)
 | |
|         # Try a monster.  This is quadratic-time in protos 0 & 1, so don't
 | |
|         # bother with those.
 | |
|         nbase = int("deadbeeffeedface", 16)
 | |
|         nbase += nbase << 1000000
 | |
|         for n in nbase, -nbase:
 | |
|             p = self.dumps(n, 2)
 | |
|             got = self.loads(p)
 | |
|             # assert_is_copy is very expensive here as it precomputes
 | |
|             # a failure message by computing the repr() of n and got,
 | |
|             # we just do the check ourselves.
 | |
|             self.assertIs(type(got), int)
 | |
|             self.assertEqual(n, got)
 | |
| 
 | |
|     def test_float(self):
 | |
|         test_values = [0.0, 4.94e-324, 1e-310, 7e-308, 6.626e-34, 0.1, 0.5,
 | |
|                        3.14, 263.44582062374053, 6.022e23, 1e30]
 | |
|         test_values = test_values + [-x for x in test_values]
 | |
|         for proto in protocols:
 | |
|             for value in test_values:
 | |
|                 pickle = self.dumps(value, proto)
 | |
|                 got = self.loads(pickle)
 | |
|                 self.assert_is_copy(value, got)
 | |
| 
 | |
|     @run_with_locale('LC_ALL', 'de_DE', 'fr_FR')
 | |
|     def test_float_format(self):
 | |
|         # make sure that floats are formatted locale independent with proto 0
 | |
|         self.assertEqual(self.dumps(1.2, 0)[0:3], b'F1.')
 | |
| 
 | |
|     def test_reduce(self):
 | |
|         for proto in protocols:
 | |
|             inst = AAA()
 | |
|             dumped = self.dumps(inst, proto)
 | |
|             loaded = self.loads(dumped)
 | |
|             self.assertEqual(loaded, REDUCE_A)
 | |
| 
 | |
|     def test_getinitargs(self):
 | |
|         for proto in protocols:
 | |
|             inst = initarg(1, 2)
 | |
|             dumped = self.dumps(inst, proto)
 | |
|             loaded = self.loads(dumped)
 | |
|             self.assert_is_copy(inst, loaded)
 | |
| 
 | |
|     def test_metaclass(self):
 | |
|         a = use_metaclass()
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(a, proto)
 | |
|             b = self.loads(s)
 | |
|             self.assertEqual(a.__class__, b.__class__)
 | |
| 
 | |
|     def test_dynamic_class(self):
 | |
|         a = create_dynamic_class("my_dynamic_class", (object,))
 | |
|         copyreg.pickle(pickling_metaclass, pickling_metaclass.__reduce__)
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(a, proto)
 | |
|             b = self.loads(s)
 | |
|             self.assertEqual(a, b)
 | |
|             self.assertIs(type(a), type(b))
 | |
| 
 | |
|     def test_structseq(self):
 | |
|         import time
 | |
|         import os
 | |
| 
 | |
|         t = time.localtime()
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(t, proto)
 | |
|             u = self.loads(s)
 | |
|             self.assert_is_copy(t, u)
 | |
|             t = os.stat(os.curdir)
 | |
|             s = self.dumps(t, proto)
 | |
|             u = self.loads(s)
 | |
|             self.assert_is_copy(t, u)
 | |
|             if hasattr(os, "statvfs"):
 | |
|                 t = os.statvfs(os.curdir)
 | |
|                 s = self.dumps(t, proto)
 | |
|                 u = self.loads(s)
 | |
|                 self.assert_is_copy(t, u)
 | |
| 
 | |
|     def test_ellipsis(self):
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(..., proto)
 | |
|             u = self.loads(s)
 | |
|             self.assertIs(..., u)
 | |
| 
 | |
|     def test_notimplemented(self):
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(NotImplemented, proto)
 | |
|             u = self.loads(s)
 | |
|             self.assertIs(NotImplemented, u)
 | |
| 
 | |
|     def test_singleton_types(self):
 | |
|         # Issue #6477: Test that types of built-in singletons can be pickled.
 | |
|         singletons = [None, ..., NotImplemented]
 | |
|         for singleton in singletons:
 | |
|             for proto in protocols:
 | |
|                 s = self.dumps(type(singleton), proto)
 | |
|                 u = self.loads(s)
 | |
|                 self.assertIs(type(singleton), u)
 | |
| 
 | |
|     # Tests for protocol 2
 | |
| 
 | |
|     def test_proto(self):
 | |
|         for proto in protocols:
 | |
|             pickled = self.dumps(None, proto)
 | |
|             if proto >= 2:
 | |
|                 proto_header = pickle.PROTO + bytes([proto])
 | |
|                 self.assertTrue(pickled.startswith(proto_header))
 | |
|             else:
 | |
|                 self.assertEqual(count_opcode(pickle.PROTO, pickled), 0)
 | |
| 
 | |
|         oob = protocols[-1] + 1     # a future protocol
 | |
|         build_none = pickle.NONE + pickle.STOP
 | |
|         badpickle = pickle.PROTO + bytes([oob]) + build_none
 | |
|         try:
 | |
|             self.loads(badpickle)
 | |
|         except ValueError as err:
 | |
|             self.assertIn("unsupported pickle protocol", str(err))
 | |
|         else:
 | |
|             self.fail("expected bad protocol number to raise ValueError")
 | |
| 
 | |
|     def test_long1(self):
 | |
|         x = 12345678910111213141516178920
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assert_is_copy(x, y)
 | |
|             self.assertEqual(opcode_in_pickle(pickle.LONG1, s), proto >= 2)
 | |
| 
 | |
|     def test_long4(self):
 | |
|         x = 12345678910111213141516178920 << (256*8)
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assert_is_copy(x, y)
 | |
|             self.assertEqual(opcode_in_pickle(pickle.LONG4, s), proto >= 2)
 | |
| 
 | |
|     def test_short_tuples(self):
 | |
|         # Map (proto, len(tuple)) to expected opcode.
 | |
|         expected_opcode = {(0, 0): pickle.TUPLE,
 | |
|                            (0, 1): pickle.TUPLE,
 | |
|                            (0, 2): pickle.TUPLE,
 | |
|                            (0, 3): pickle.TUPLE,
 | |
|                            (0, 4): pickle.TUPLE,
 | |
| 
 | |
|                            (1, 0): pickle.EMPTY_TUPLE,
 | |
|                            (1, 1): pickle.TUPLE,
 | |
|                            (1, 2): pickle.TUPLE,
 | |
|                            (1, 3): pickle.TUPLE,
 | |
|                            (1, 4): pickle.TUPLE,
 | |
| 
 | |
|                            (2, 0): pickle.EMPTY_TUPLE,
 | |
|                            (2, 1): pickle.TUPLE1,
 | |
|                            (2, 2): pickle.TUPLE2,
 | |
|                            (2, 3): pickle.TUPLE3,
 | |
|                            (2, 4): pickle.TUPLE,
 | |
| 
 | |
|                            (3, 0): pickle.EMPTY_TUPLE,
 | |
|                            (3, 1): pickle.TUPLE1,
 | |
|                            (3, 2): pickle.TUPLE2,
 | |
|                            (3, 3): pickle.TUPLE3,
 | |
|                            (3, 4): pickle.TUPLE,
 | |
|                           }
 | |
|         a = ()
 | |
|         b = (1,)
 | |
|         c = (1, 2)
 | |
|         d = (1, 2, 3)
 | |
|         e = (1, 2, 3, 4)
 | |
|         for proto in protocols:
 | |
|             for x in a, b, c, d, e:
 | |
|                 s = self.dumps(x, proto)
 | |
|                 y = self.loads(s)
 | |
|                 self.assert_is_copy(x, y)
 | |
|                 expected = expected_opcode[min(proto, 3), len(x)]
 | |
|                 self.assertTrue(opcode_in_pickle(expected, s))
 | |
| 
 | |
|     def test_singletons(self):
 | |
|         # Map (proto, singleton) to expected opcode.
 | |
|         expected_opcode = {(0, None): pickle.NONE,
 | |
|                            (1, None): pickle.NONE,
 | |
|                            (2, None): pickle.NONE,
 | |
|                            (3, None): pickle.NONE,
 | |
| 
 | |
|                            (0, True): pickle.INT,
 | |
|                            (1, True): pickle.INT,
 | |
|                            (2, True): pickle.NEWTRUE,
 | |
|                            (3, True): pickle.NEWTRUE,
 | |
| 
 | |
|                            (0, False): pickle.INT,
 | |
|                            (1, False): pickle.INT,
 | |
|                            (2, False): pickle.NEWFALSE,
 | |
|                            (3, False): pickle.NEWFALSE,
 | |
|                           }
 | |
|         for proto in protocols:
 | |
|             for x in None, False, True:
 | |
|                 s = self.dumps(x, proto)
 | |
|                 y = self.loads(s)
 | |
|                 self.assertTrue(x is y, (proto, x, s, y))
 | |
|                 expected = expected_opcode[min(proto, 3), x]
 | |
|                 self.assertTrue(opcode_in_pickle(expected, s))
 | |
| 
 | |
|     def test_newobj_tuple(self):
 | |
|         x = MyTuple([1, 2, 3])
 | |
|         x.foo = 42
 | |
|         x.bar = "hello"
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assert_is_copy(x, y)
 | |
| 
 | |
|     def test_newobj_list(self):
 | |
|         x = MyList([1, 2, 3])
 | |
|         x.foo = 42
 | |
|         x.bar = "hello"
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assert_is_copy(x, y)
 | |
| 
 | |
|     def test_newobj_generic(self):
 | |
|         for proto in protocols:
 | |
|             for C in myclasses:
 | |
|                 B = C.__base__
 | |
|                 x = C(C.sample)
 | |
|                 x.foo = 42
 | |
|                 s = self.dumps(x, proto)
 | |
|                 y = self.loads(s)
 | |
|                 detail = (proto, C, B, x, y, type(y))
 | |
|                 self.assert_is_copy(x, y) # XXX revisit
 | |
|                 self.assertEqual(B(x), B(y), detail)
 | |
|                 self.assertEqual(x.__dict__, y.__dict__, detail)
 | |
| 
 | |
|     def test_newobj_proxies(self):
 | |
|         # NEWOBJ should use the __class__ rather than the raw type
 | |
|         classes = myclasses[:]
 | |
|         # Cannot create weakproxies to these classes
 | |
|         for c in (MyInt, MyTuple):
 | |
|             classes.remove(c)
 | |
|         for proto in protocols:
 | |
|             for C in classes:
 | |
|                 B = C.__base__
 | |
|                 x = C(C.sample)
 | |
|                 x.foo = 42
 | |
|                 p = weakref.proxy(x)
 | |
|                 s = self.dumps(p, proto)
 | |
|                 y = self.loads(s)
 | |
|                 self.assertEqual(type(y), type(x))  # rather than type(p)
 | |
|                 detail = (proto, C, B, x, y, type(y))
 | |
|                 self.assertEqual(B(x), B(y), detail)
 | |
|                 self.assertEqual(x.__dict__, y.__dict__, detail)
 | |
| 
 | |
|     def test_newobj_overridden_new(self):
 | |
|         # Test that Python class with C implemented __new__ is pickleable
 | |
|         for proto in protocols:
 | |
|             x = MyIntWithNew2(1)
 | |
|             x.foo = 42
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assertIs(type(y), MyIntWithNew2)
 | |
|             self.assertEqual(int(y), 1)
 | |
|             self.assertEqual(y.foo, 42)
 | |
| 
 | |
|     def test_newobj_not_class(self):
 | |
|         # Issue 24552
 | |
|         global SimpleNewObj
 | |
|         save = SimpleNewObj
 | |
|         o = SimpleNewObj.__new__(SimpleNewObj)
 | |
|         b = self.dumps(o, 4)
 | |
|         try:
 | |
|             SimpleNewObj = 42
 | |
|             self.assertRaises((TypeError, pickle.UnpicklingError), self.loads, b)
 | |
|         finally:
 | |
|             SimpleNewObj = save
 | |
| 
 | |
|     # Register a type with copyreg, with extension code extcode.  Pickle
 | |
|     # an object of that type.  Check that the resulting pickle uses opcode
 | |
|     # (EXT[124]) under proto 2, and not in proto 1.
 | |
| 
 | |
|     def produce_global_ext(self, extcode, opcode):
 | |
|         e = ExtensionSaver(extcode)
 | |
|         try:
 | |
|             copyreg.add_extension(__name__, "MyList", extcode)
 | |
|             x = MyList([1, 2, 3])
 | |
|             x.foo = 42
 | |
|             x.bar = "hello"
 | |
| 
 | |
|             # Dump using protocol 1 for comparison.
 | |
|             s1 = self.dumps(x, 1)
 | |
|             self.assertIn(__name__.encode("utf-8"), s1)
 | |
|             self.assertIn(b"MyList", s1)
 | |
|             self.assertFalse(opcode_in_pickle(opcode, s1))
 | |
| 
 | |
|             y = self.loads(s1)
 | |
|             self.assert_is_copy(x, y)
 | |
| 
 | |
|             # Dump using protocol 2 for test.
 | |
|             s2 = self.dumps(x, 2)
 | |
|             self.assertNotIn(__name__.encode("utf-8"), s2)
 | |
|             self.assertNotIn(b"MyList", s2)
 | |
|             self.assertEqual(opcode_in_pickle(opcode, s2), True, repr(s2))
 | |
| 
 | |
|             y = self.loads(s2)
 | |
|             self.assert_is_copy(x, y)
 | |
|         finally:
 | |
|             e.restore()
 | |
| 
 | |
|     def test_global_ext1(self):
 | |
|         self.produce_global_ext(0x00000001, pickle.EXT1)  # smallest EXT1 code
 | |
|         self.produce_global_ext(0x000000ff, pickle.EXT1)  # largest EXT1 code
 | |
| 
 | |
|     def test_global_ext2(self):
 | |
|         self.produce_global_ext(0x00000100, pickle.EXT2)  # smallest EXT2 code
 | |
|         self.produce_global_ext(0x0000ffff, pickle.EXT2)  # largest EXT2 code
 | |
|         self.produce_global_ext(0x0000abcd, pickle.EXT2)  # check endianness
 | |
| 
 | |
|     def test_global_ext4(self):
 | |
|         self.produce_global_ext(0x00010000, pickle.EXT4)  # smallest EXT4 code
 | |
|         self.produce_global_ext(0x7fffffff, pickle.EXT4)  # largest EXT4 code
 | |
|         self.produce_global_ext(0x12abcdef, pickle.EXT4)  # check endianness
 | |
| 
 | |
|     def test_list_chunking(self):
 | |
|         n = 10  # too small to chunk
 | |
|         x = list(range(n))
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assert_is_copy(x, y)
 | |
|             num_appends = count_opcode(pickle.APPENDS, s)
 | |
|             self.assertEqual(num_appends, proto > 0)
 | |
| 
 | |
|         n = 2500  # expect at least two chunks when proto > 0
 | |
|         x = list(range(n))
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assert_is_copy(x, y)
 | |
|             num_appends = count_opcode(pickle.APPENDS, s)
 | |
|             if proto == 0:
 | |
|                 self.assertEqual(num_appends, 0)
 | |
|             else:
 | |
|                 self.assertTrue(num_appends >= 2)
 | |
| 
 | |
|     def test_dict_chunking(self):
 | |
|         n = 10  # too small to chunk
 | |
|         x = dict.fromkeys(range(n))
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(x, proto)
 | |
|             self.assertIsInstance(s, bytes_types)
 | |
|             y = self.loads(s)
 | |
|             self.assert_is_copy(x, y)
 | |
|             num_setitems = count_opcode(pickle.SETITEMS, s)
 | |
|             self.assertEqual(num_setitems, proto > 0)
 | |
| 
 | |
|         n = 2500  # expect at least two chunks when proto > 0
 | |
|         x = dict.fromkeys(range(n))
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assert_is_copy(x, y)
 | |
|             num_setitems = count_opcode(pickle.SETITEMS, s)
 | |
|             if proto == 0:
 | |
|                 self.assertEqual(num_setitems, 0)
 | |
|             else:
 | |
|                 self.assertTrue(num_setitems >= 2)
 | |
| 
 | |
|     def test_set_chunking(self):
 | |
|         n = 10  # too small to chunk
 | |
|         x = set(range(n))
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assert_is_copy(x, y)
 | |
|             num_additems = count_opcode(pickle.ADDITEMS, s)
 | |
|             if proto < 4:
 | |
|                 self.assertEqual(num_additems, 0)
 | |
|             else:
 | |
|                 self.assertEqual(num_additems, 1)
 | |
| 
 | |
|         n = 2500  # expect at least two chunks when proto >= 4
 | |
|         x = set(range(n))
 | |
|         for proto in protocols:
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assert_is_copy(x, y)
 | |
|             num_additems = count_opcode(pickle.ADDITEMS, s)
 | |
|             if proto < 4:
 | |
|                 self.assertEqual(num_additems, 0)
 | |
|             else:
 | |
|                 self.assertGreaterEqual(num_additems, 2)
 | |
| 
 | |
|     def test_simple_newobj(self):
 | |
|         x = SimpleNewObj.__new__(SimpleNewObj, 0xface)  # avoid __init__
 | |
|         x.abc = 666
 | |
|         for proto in protocols:
 | |
|             with self.subTest(proto=proto):
 | |
|                 s = self.dumps(x, proto)
 | |
|                 if proto < 1:
 | |
|                     self.assertIn(b'\nI64206', s)  # INT
 | |
|                 else:
 | |
|                     self.assertIn(b'M\xce\xfa', s)  # BININT2
 | |
|                 self.assertEqual(opcode_in_pickle(pickle.NEWOBJ, s),
 | |
|                                  2 <= proto)
 | |
|                 self.assertFalse(opcode_in_pickle(pickle.NEWOBJ_EX, s))
 | |
|                 y = self.loads(s)   # will raise TypeError if __init__ called
 | |
|                 self.assert_is_copy(x, y)
 | |
| 
 | |
|     def test_complex_newobj(self):
 | |
|         x = ComplexNewObj.__new__(ComplexNewObj, 0xface)  # avoid __init__
 | |
|         x.abc = 666
 | |
|         for proto in protocols:
 | |
|             with self.subTest(proto=proto):
 | |
|                 s = self.dumps(x, proto)
 | |
|                 if proto < 1:
 | |
|                     self.assertIn(b'\nI64206', s)  # INT
 | |
|                 elif proto < 2:
 | |
|                     self.assertIn(b'M\xce\xfa', s)  # BININT2
 | |
|                 elif proto < 4:
 | |
|                     self.assertIn(b'X\x04\x00\x00\x00FACE', s)  # BINUNICODE
 | |
|                 else:
 | |
|                     self.assertIn(b'\x8c\x04FACE', s)  # SHORT_BINUNICODE
 | |
|                 self.assertEqual(opcode_in_pickle(pickle.NEWOBJ, s),
 | |
|                                  2 <= proto)
 | |
|                 self.assertFalse(opcode_in_pickle(pickle.NEWOBJ_EX, s))
 | |
|                 y = self.loads(s)   # will raise TypeError if __init__ called
 | |
|                 self.assert_is_copy(x, y)
 | |
| 
 | |
|     def test_complex_newobj_ex(self):
 | |
|         x = ComplexNewObjEx.__new__(ComplexNewObjEx, 0xface)  # avoid __init__
 | |
|         x.abc = 666
 | |
|         for proto in protocols:
 | |
|             with self.subTest(proto=proto):
 | |
|                 s = self.dumps(x, proto)
 | |
|                 if proto < 1:
 | |
|                     self.assertIn(b'\nI64206', s)  # INT
 | |
|                 elif proto < 2:
 | |
|                     self.assertIn(b'M\xce\xfa', s)  # BININT2
 | |
|                 elif proto < 4:
 | |
|                     self.assertIn(b'X\x04\x00\x00\x00FACE', s)  # BINUNICODE
 | |
|                 else:
 | |
|                     self.assertIn(b'\x8c\x04FACE', s)  # SHORT_BINUNICODE
 | |
|                 self.assertFalse(opcode_in_pickle(pickle.NEWOBJ, s))
 | |
|                 self.assertEqual(opcode_in_pickle(pickle.NEWOBJ_EX, s),
 | |
|                                  4 <= proto)
 | |
|                 y = self.loads(s)   # will raise TypeError if __init__ called
 | |
|                 self.assert_is_copy(x, y)
 | |
| 
 | |
|     def test_newobj_list_slots(self):
 | |
|         x = SlotList([1, 2, 3])
 | |
|         x.foo = 42
 | |
|         x.bar = "hello"
 | |
|         s = self.dumps(x, 2)
 | |
|         y = self.loads(s)
 | |
|         self.assert_is_copy(x, y)
 | |
| 
 | |
|     def test_reduce_overrides_default_reduce_ex(self):
 | |
|         for proto in protocols:
 | |
|             x = REX_one()
 | |
|             self.assertEqual(x._reduce_called, 0)
 | |
|             s = self.dumps(x, proto)
 | |
|             self.assertEqual(x._reduce_called, 1)
 | |
|             y = self.loads(s)
 | |
|             self.assertEqual(y._reduce_called, 0)
 | |
| 
 | |
|     def test_reduce_ex_called(self):
 | |
|         for proto in protocols:
 | |
|             x = REX_two()
 | |
|             self.assertEqual(x._proto, None)
 | |
|             s = self.dumps(x, proto)
 | |
|             self.assertEqual(x._proto, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assertEqual(y._proto, None)
 | |
| 
 | |
|     def test_reduce_ex_overrides_reduce(self):
 | |
|         for proto in protocols:
 | |
|             x = REX_three()
 | |
|             self.assertEqual(x._proto, None)
 | |
|             s = self.dumps(x, proto)
 | |
|             self.assertEqual(x._proto, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assertEqual(y._proto, None)
 | |
| 
 | |
|     def test_reduce_ex_calls_base(self):
 | |
|         for proto in protocols:
 | |
|             x = REX_four()
 | |
|             self.assertEqual(x._proto, None)
 | |
|             s = self.dumps(x, proto)
 | |
|             self.assertEqual(x._proto, proto)
 | |
|             y = self.loads(s)
 | |
|             self.assertEqual(y._proto, proto)
 | |
| 
 | |
|     def test_reduce_calls_base(self):
 | |
|         for proto in protocols:
 | |
|             x = REX_five()
 | |
|             self.assertEqual(x._reduce_called, 0)
 | |
|             s = self.dumps(x, proto)
 | |
|             self.assertEqual(x._reduce_called, 1)
 | |
|             y = self.loads(s)
 | |
|             self.assertEqual(y._reduce_called, 1)
 | |
| 
 | |
|     @no_tracing
 | |
|     def test_bad_getattr(self):
 | |
|         # Issue #3514: crash when there is an infinite loop in __getattr__
 | |
|         x = BadGetattr()
 | |
|         for proto in protocols:
 | |
|             self.assertRaises(RuntimeError, self.dumps, x, proto)
 | |
| 
 | |
|     def test_reduce_bad_iterator(self):
 | |
|         # Issue4176: crash when 4th and 5th items of __reduce__()
 | |
|         # are not iterators
 | |
|         class C(object):
 | |
|             def __reduce__(self):
 | |
|                 # 4th item is not an iterator
 | |
|                 return list, (), None, [], None
 | |
|         class D(object):
 | |
|             def __reduce__(self):
 | |
|                 # 5th item is not an iterator
 | |
|                 return dict, (), None, None, []
 | |
| 
 | |
|         # Python implementation is less strict and also accepts iterables.
 | |
|         for proto in protocols:
 | |
|             try:
 | |
|                 self.dumps(C(), proto)
 | |
|             except pickle.PicklingError:
 | |
|                 pass
 | |
|             try:
 | |
|                 self.dumps(D(), proto)
 | |
|             except pickle.PicklingError:
 | |
|                 pass
 | |
| 
 | |
|     def test_many_puts_and_gets(self):
 | |
|         # Test that internal data structures correctly deal with lots of
 | |
|         # puts/gets.
 | |
|         keys = ("aaa" + str(i) for i in range(100))
 | |
|         large_dict = dict((k, [4, 5, 6]) for k in keys)
 | |
|         obj = [dict(large_dict), dict(large_dict), dict(large_dict)]
 | |
| 
 | |
|         for proto in protocols:
 | |
|             with self.subTest(proto=proto):
 | |
|                 dumped = self.dumps(obj, proto)
 | |
|                 loaded = self.loads(dumped)
 | |
|                 self.assert_is_copy(obj, loaded)
 | |
| 
 | |
|     def test_attribute_name_interning(self):
 | |
|         # Test that attribute names of pickled objects are interned when
 | |
|         # unpickling.
 | |
|         for proto in protocols:
 | |
|             x = C()
 | |
|             x.foo = 42
 | |
|             x.bar = "hello"
 | |
|             s = self.dumps(x, proto)
 | |
|             y = self.loads(s)
 | |
|             x_keys = sorted(x.__dict__)
 | |
|             y_keys = sorted(y.__dict__)
 | |
|             for x_key, y_key in zip(x_keys, y_keys):
 | |
|                 self.assertIs(x_key, y_key)
 | |
| 
 | |
|     def test_pickle_to_2x(self):
 | |
|         # Pickle non-trivial data with protocol 2, expecting that it yields
 | |
|         # the same result as Python 2.x did.
 | |
|         # NOTE: this test is a bit too strong since we can produce different
 | |
|         # bytecode that 2.x will still understand.
 | |
|         dumped = self.dumps(range(5), 2)
 | |
|         self.assertEqual(dumped, DATA_XRANGE)
 | |
|         dumped = self.dumps(set([3]), 2)
 | |
|         self.assertEqual(dumped, DATA_SET2)
 | |
| 
 | |
|     def test_large_pickles(self):
 | |
|         # Test the correctness of internal buffering routines when handling
 | |
|         # large data.
 | |
|         for proto in protocols:
 | |
|             data = (1, min, b'xy' * (30 * 1024), len)
 | |
|             dumped = self.dumps(data, proto)
 | |
|             loaded = self.loads(dumped)
 | |
|             self.assertEqual(len(loaded), len(data))
 | |
|             self.assertEqual(loaded, data)
 | |
| 
 | |
|     def test_int_pickling_efficiency(self):
 | |
|         # Test compacity of int representation (see issue #12744)
 | |
|         for proto in protocols:
 | |
|             with self.subTest(proto=proto):
 | |
|                 pickles = [self.dumps(2**n, proto) for n in range(70)]
 | |
|                 sizes = list(map(len, pickles))
 | |
|                 # the size function is monotonic
 | |
|                 self.assertEqual(sorted(sizes), sizes)
 | |
|                 if proto >= 2:
 | |
|                     for p in pickles:
 | |
|                         self.assertFalse(opcode_in_pickle(pickle.LONG, p))
 | |
| 
 | |
|     def _check_pickling_with_opcode(self, obj, opcode, proto):
 | |
|         pickled = self.dumps(obj, proto)
 | |
|         self.assertTrue(opcode_in_pickle(opcode, pickled))
 | |
|         unpickled = self.loads(pickled)
 | |
|         self.assertEqual(obj, unpickled)
 | |
| 
 | |
|     def test_appends_on_non_lists(self):
 | |
|         # Issue #17720
 | |
|         obj = REX_six([1, 2, 3])
 | |
|         for proto in protocols:
 | |
|             if proto == 0:
 | |
|                 self._check_pickling_with_opcode(obj, pickle.APPEND, proto)
 | |
|             else:
 | |
|                 self._check_pickling_with_opcode(obj, pickle.APPENDS, proto)
 | |
| 
 | |
|     def test_setitems_on_non_dicts(self):
 | |
|         obj = REX_seven({1: -1, 2: -2, 3: -3})
 | |
|         for proto in protocols:
 | |
|             if proto == 0:
 | |
|                 self._check_pickling_with_opcode(obj, pickle.SETITEM, proto)
 | |
|             else:
 | |
|                 self._check_pickling_with_opcode(obj, pickle.SETITEMS, proto)
 | |
| 
 | |
|     # Exercise framing (proto >= 4) for significant workloads
 | |
| 
 | |
|     FRAME_SIZE_MIN = 4
 | |
|     FRAME_SIZE_TARGET = 64 * 1024
 | |
| 
 | |
|     def check_frame_opcodes(self, pickled):
 | |
|         """
 | |
|         Check the arguments of FRAME opcodes in a protocol 4+ pickle.
 | |
| 
 | |
|         Note that binary objects that are larger than FRAME_SIZE_TARGET are not
 | |
|         framed by default and are therefore considered a frame by themselves in
 | |
|         the following consistency check.
 | |
|         """
 | |
|         frame_end = frameless_start = None
 | |
|         frameless_opcodes = {'BINBYTES', 'BINUNICODE', 'BINBYTES8',
 | |
|                              'BINUNICODE8', 'BYTEARRAY8'}
 | |
|         for op, arg, pos in pickletools.genops(pickled):
 | |
|             if frame_end is not None:
 | |
|                 self.assertLessEqual(pos, frame_end)
 | |
|                 if pos == frame_end:
 | |
|                     frame_end = None
 | |
| 
 | |
|             if frame_end is not None:  # framed
 | |
|                 self.assertNotEqual(op.name, 'FRAME')
 | |
|                 if op.name in frameless_opcodes:
 | |
|                     # Only short bytes and str objects should be written
 | |
|                     # in a frame
 | |
|                     self.assertLessEqual(len(arg), self.FRAME_SIZE_TARGET)
 | |
| 
 | |
|             else:  # not framed
 | |
|                 if (op.name == 'FRAME' or
 | |
|                     (op.name in frameless_opcodes and
 | |
|                      len(arg) > self.FRAME_SIZE_TARGET)):
 | |
|                     # Frame or large bytes or str object
 | |
|                     if frameless_start is not None:
 | |
|                         # Only short data should be written outside of a frame
 | |
|                         self.assertLess(pos - frameless_start,
 | |
|                                         self.FRAME_SIZE_MIN)
 | |
|                         frameless_start = None
 | |
|                 elif frameless_start is None and op.name != 'PROTO':
 | |
|                     frameless_start = pos
 | |
| 
 | |
|             if op.name == 'FRAME':
 | |
|                 self.assertGreaterEqual(arg, self.FRAME_SIZE_MIN)
 | |
|                 frame_end = pos + 9 + arg
 | |
| 
 | |
|         pos = len(pickled)
 | |
|         if frame_end is not None:
 | |
|             self.assertEqual(frame_end, pos)
 | |
|         elif frameless_start is not None:
 | |
|             self.assertLess(pos - frameless_start, self.FRAME_SIZE_MIN)
 | |
| 
 | |
|     @support.skip_if_pgo_task
 | |
|     def test_framing_many_objects(self):
 | |
|         obj = list(range(10**5))
 | |
|         for proto in range(4, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             with self.subTest(proto=proto):
 | |
|                 pickled = self.dumps(obj, proto)
 | |
|                 unpickled = self.loads(pickled)
 | |
|                 self.assertEqual(obj, unpickled)
 | |
|                 bytes_per_frame = (len(pickled) /
 | |
|                                    count_opcode(pickle.FRAME, pickled))
 | |
|                 self.assertGreater(bytes_per_frame,
 | |
|                                    self.FRAME_SIZE_TARGET / 2)
 | |
|                 self.assertLessEqual(bytes_per_frame,
 | |
|                                      self.FRAME_SIZE_TARGET * 1)
 | |
|                 self.check_frame_opcodes(pickled)
 | |
| 
 | |
|     def test_framing_large_objects(self):
 | |
|         N = 1024 * 1024
 | |
|         small_items = [[i] for i in range(10)]
 | |
|         obj = [b'x' * N, *small_items, b'y' * N, 'z' * N]
 | |
|         for proto in range(4, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             for fast in [False, True]:
 | |
|                 with self.subTest(proto=proto, fast=fast):
 | |
|                     if not fast:
 | |
|                         # fast=False by default.
 | |
|                         # This covers in-memory pickling with pickle.dumps().
 | |
|                         pickled = self.dumps(obj, proto)
 | |
|                     else:
 | |
|                         # Pickler is required when fast=True.
 | |
|                         if not hasattr(self, 'pickler'):
 | |
|                             continue
 | |
|                         buf = io.BytesIO()
 | |
|                         pickler = self.pickler(buf, protocol=proto)
 | |
|                         pickler.fast = fast
 | |
|                         pickler.dump(obj)
 | |
|                         pickled = buf.getvalue()
 | |
|                     unpickled = self.loads(pickled)
 | |
|                     # More informative error message in case of failure.
 | |
|                     self.assertEqual([len(x) for x in obj],
 | |
|                                      [len(x) for x in unpickled])
 | |
|                     # Perform full equality check if the lengths match.
 | |
|                     self.assertEqual(obj, unpickled)
 | |
|                     n_frames = count_opcode(pickle.FRAME, pickled)
 | |
|                     # A single frame for small objects between
 | |
|                     # first two large objects.
 | |
|                     self.assertEqual(n_frames, 1)
 | |
|                     self.check_frame_opcodes(pickled)
 | |
| 
 | |
|     def test_optional_frames(self):
 | |
|         if pickle.HIGHEST_PROTOCOL < 4:
 | |
|             return
 | |
| 
 | |
|         def remove_frames(pickled, keep_frame=None):
 | |
|             """Remove frame opcodes from the given pickle."""
 | |
|             frame_starts = []
 | |
|             # 1 byte for the opcode and 8 for the argument
 | |
|             frame_opcode_size = 9
 | |
|             for opcode, _, pos in pickletools.genops(pickled):
 | |
|                 if opcode.name == 'FRAME':
 | |
|                     frame_starts.append(pos)
 | |
| 
 | |
|             newpickle = bytearray()
 | |
|             last_frame_end = 0
 | |
|             for i, pos in enumerate(frame_starts):
 | |
|                 if keep_frame and keep_frame(i):
 | |
|                     continue
 | |
|                 newpickle += pickled[last_frame_end:pos]
 | |
|                 last_frame_end = pos + frame_opcode_size
 | |
|             newpickle += pickled[last_frame_end:]
 | |
|             return newpickle
 | |
| 
 | |
|         frame_size = self.FRAME_SIZE_TARGET
 | |
|         num_frames = 20
 | |
|         # Large byte objects (dict values) intermittent with small objects
 | |
|         # (dict keys)
 | |
|         for bytes_type in (bytes, bytearray):
 | |
|             obj = {i: bytes_type([i]) * frame_size for i in range(num_frames)}
 | |
| 
 | |
|             for proto in range(4, pickle.HIGHEST_PROTOCOL + 1):
 | |
|                 pickled = self.dumps(obj, proto)
 | |
| 
 | |
|                 frameless_pickle = remove_frames(pickled)
 | |
|                 self.assertEqual(count_opcode(pickle.FRAME, frameless_pickle), 0)
 | |
|                 self.assertEqual(obj, self.loads(frameless_pickle))
 | |
| 
 | |
|                 some_frames_pickle = remove_frames(pickled, lambda i: i % 2)
 | |
|                 self.assertLess(count_opcode(pickle.FRAME, some_frames_pickle),
 | |
|                                 count_opcode(pickle.FRAME, pickled))
 | |
|                 self.assertEqual(obj, self.loads(some_frames_pickle))
 | |
| 
 | |
|     @support.skip_if_pgo_task
 | |
|     def test_framed_write_sizes_with_delayed_writer(self):
 | |
|         class ChunkAccumulator:
 | |
|             """Accumulate pickler output in a list of raw chunks."""
 | |
|             def __init__(self):
 | |
|                 self.chunks = []
 | |
|             def write(self, chunk):
 | |
|                 self.chunks.append(chunk)
 | |
|             def concatenate_chunks(self):
 | |
|                 return b"".join(self.chunks)
 | |
| 
 | |
|         for proto in range(4, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             objects = [(str(i).encode('ascii'), i % 42, {'i': str(i)})
 | |
|                        for i in range(int(1e4))]
 | |
|             # Add a large unique ASCII string
 | |
|             objects.append('0123456789abcdef' *
 | |
|                            (self.FRAME_SIZE_TARGET // 16 + 1))
 | |
| 
 | |
|             # Protocol 4 packs groups of small objects into frames and issues
 | |
|             # calls to write only once or twice per frame:
 | |
|             # The C pickler issues one call to write per-frame (header and
 | |
|             # contents) while Python pickler issues two calls to write: one for
 | |
|             # the frame header and one for the frame binary contents.
 | |
|             writer = ChunkAccumulator()
 | |
|             self.pickler(writer, proto).dump(objects)
 | |
| 
 | |
|             # Actually read the binary content of the chunks after the end
 | |
|             # of the call to dump: any memoryview passed to write should not
 | |
|             # be released otherwise this delayed access would not be possible.
 | |
|             pickled = writer.concatenate_chunks()
 | |
|             reconstructed = self.loads(pickled)
 | |
|             self.assertEqual(reconstructed, objects)
 | |
|             self.assertGreater(len(writer.chunks), 1)
 | |
| 
 | |
|             # memoryviews should own the memory.
 | |
|             del objects
 | |
|             support.gc_collect()
 | |
|             self.assertEqual(writer.concatenate_chunks(), pickled)
 | |
| 
 | |
|             n_frames = (len(pickled) - 1) // self.FRAME_SIZE_TARGET + 1
 | |
|             # There should be at least one call to write per frame
 | |
|             self.assertGreaterEqual(len(writer.chunks), n_frames)
 | |
| 
 | |
|             # but not too many either: there can be one for the proto,
 | |
|             # one per-frame header, one per frame for the actual contents,
 | |
|             # and two for the header.
 | |
|             self.assertLessEqual(len(writer.chunks), 2 * n_frames + 3)
 | |
| 
 | |
|             chunk_sizes = [len(c) for c in writer.chunks]
 | |
|             large_sizes = [s for s in chunk_sizes
 | |
|                            if s >= self.FRAME_SIZE_TARGET]
 | |
|             medium_sizes = [s for s in chunk_sizes
 | |
|                            if 9 < s < self.FRAME_SIZE_TARGET]
 | |
|             small_sizes = [s for s in chunk_sizes if s <= 9]
 | |
| 
 | |
|             # Large chunks should not be too large:
 | |
|             for chunk_size in large_sizes:
 | |
|                 self.assertLess(chunk_size, 2 * self.FRAME_SIZE_TARGET,
 | |
|                                 chunk_sizes)
 | |
|             # There shouldn't bee too many small chunks: the protocol header,
 | |
|             # the frame headers and the large string headers are written
 | |
|             # in small chunks.
 | |
|             self.assertLessEqual(len(small_sizes),
 | |
|                                  len(large_sizes) + len(medium_sizes) + 3,
 | |
|                                  chunk_sizes)
 | |
| 
 | |
|     def test_nested_names(self):
 | |
|         global Nested
 | |
|         class Nested:
 | |
|             class A:
 | |
|                 class B:
 | |
|                     class C:
 | |
|                         pass
 | |
|         for proto in range(pickle.HIGHEST_PROTOCOL + 1):
 | |
|             for obj in [Nested.A, Nested.A.B, Nested.A.B.C]:
 | |
|                 with self.subTest(proto=proto, obj=obj):
 | |
|                     unpickled = self.loads(self.dumps(obj, proto))
 | |
|                     self.assertIs(obj, unpickled)
 | |
| 
 | |
|     def test_recursive_nested_names(self):
 | |
|         global Recursive
 | |
|         class Recursive:
 | |
|             pass
 | |
|         Recursive.mod = sys.modules[Recursive.__module__]
 | |
|         Recursive.__qualname__ = 'Recursive.mod.Recursive'
 | |
|         for proto in range(pickle.HIGHEST_PROTOCOL + 1):
 | |
|             with self.subTest(proto=proto):
 | |
|                 unpickled = self.loads(self.dumps(Recursive, proto))
 | |
|                 self.assertIs(unpickled, Recursive)
 | |
|         del Recursive.mod # break reference loop
 | |
| 
 | |
|     def test_py_methods(self):
 | |
|         global PyMethodsTest
 | |
|         class PyMethodsTest:
 | |
|             @staticmethod
 | |
|             def cheese():
 | |
|                 return "cheese"
 | |
|             @classmethod
 | |
|             def wine(cls):
 | |
|                 assert cls is PyMethodsTest
 | |
|                 return "wine"
 | |
|             def biscuits(self):
 | |
|                 assert isinstance(self, PyMethodsTest)
 | |
|                 return "biscuits"
 | |
|             class Nested:
 | |
|                 "Nested class"
 | |
|                 @staticmethod
 | |
|                 def ketchup():
 | |
|                     return "ketchup"
 | |
|                 @classmethod
 | |
|                 def maple(cls):
 | |
|                     assert cls is PyMethodsTest.Nested
 | |
|                     return "maple"
 | |
|                 def pie(self):
 | |
|                     assert isinstance(self, PyMethodsTest.Nested)
 | |
|                     return "pie"
 | |
| 
 | |
|         py_methods = (
 | |
|             PyMethodsTest.cheese,
 | |
|             PyMethodsTest.wine,
 | |
|             PyMethodsTest().biscuits,
 | |
|             PyMethodsTest.Nested.ketchup,
 | |
|             PyMethodsTest.Nested.maple,
 | |
|             PyMethodsTest.Nested().pie
 | |
|         )
 | |
|         py_unbound_methods = (
 | |
|             (PyMethodsTest.biscuits, PyMethodsTest),
 | |
|             (PyMethodsTest.Nested.pie, PyMethodsTest.Nested)
 | |
|         )
 | |
|         for proto in range(pickle.HIGHEST_PROTOCOL + 1):
 | |
|             for method in py_methods:
 | |
|                 with self.subTest(proto=proto, method=method):
 | |
|                     unpickled = self.loads(self.dumps(method, proto))
 | |
|                     self.assertEqual(method(), unpickled())
 | |
|             for method, cls in py_unbound_methods:
 | |
|                 obj = cls()
 | |
|                 with self.subTest(proto=proto, method=method):
 | |
|                     unpickled = self.loads(self.dumps(method, proto))
 | |
|                     self.assertEqual(method(obj), unpickled(obj))
 | |
| 
 | |
|     def test_c_methods(self):
 | |
|         global Subclass
 | |
|         class Subclass(tuple):
 | |
|             class Nested(str):
 | |
|                 pass
 | |
| 
 | |
|         c_methods = (
 | |
|             # bound built-in method
 | |
|             ("abcd".index, ("c",)),
 | |
|             # unbound built-in method
 | |
|             (str.index, ("abcd", "c")),
 | |
|             # bound "slot" method
 | |
|             ([1, 2, 3].__len__, ()),
 | |
|             # unbound "slot" method
 | |
|             (list.__len__, ([1, 2, 3],)),
 | |
|             # bound "coexist" method
 | |
|             ({1, 2}.__contains__, (2,)),
 | |
|             # unbound "coexist" method
 | |
|             (set.__contains__, ({1, 2}, 2)),
 | |
|             # built-in class method
 | |
|             (dict.fromkeys, (("a", 1), ("b", 2))),
 | |
|             # built-in static method
 | |
|             (bytearray.maketrans, (b"abc", b"xyz")),
 | |
|             # subclass methods
 | |
|             (Subclass([1,2,2]).count, (2,)),
 | |
|             (Subclass.count, (Subclass([1,2,2]), 2)),
 | |
|             (Subclass.Nested("sweet").count, ("e",)),
 | |
|             (Subclass.Nested.count, (Subclass.Nested("sweet"), "e")),
 | |
|         )
 | |
|         for proto in range(pickle.HIGHEST_PROTOCOL + 1):
 | |
|             for method, args in c_methods:
 | |
|                 with self.subTest(proto=proto, method=method):
 | |
|                     unpickled = self.loads(self.dumps(method, proto))
 | |
|                     self.assertEqual(method(*args), unpickled(*args))
 | |
| 
 | |
|     def test_compat_pickle(self):
 | |
|         tests = [
 | |
|             (range(1, 7), '__builtin__', 'xrange'),
 | |
|             (map(int, '123'), 'itertools', 'imap'),
 | |
|             (functools.reduce, '__builtin__', 'reduce'),
 | |
|             (dbm.whichdb, 'whichdb', 'whichdb'),
 | |
|             (Exception(), 'exceptions', 'Exception'),
 | |
|             (collections.UserDict(), 'UserDict', 'IterableUserDict'),
 | |
|             (collections.UserList(), 'UserList', 'UserList'),
 | |
|             (collections.defaultdict(), 'collections', 'defaultdict'),
 | |
|         ]
 | |
|         for val, mod, name in tests:
 | |
|             for proto in range(3):
 | |
|                 with self.subTest(type=type(val), proto=proto):
 | |
|                     pickled = self.dumps(val, proto)
 | |
|                     self.assertIn(('c%s\n%s' % (mod, name)).encode(), pickled)
 | |
|                     self.assertIs(type(self.loads(pickled)), type(val))
 | |
| 
 | |
|     def test_local_lookup_error(self):
 | |
|         # Test that whichmodule() errors out cleanly when looking up
 | |
|         # an assumed globally-reachable object fails.
 | |
|         def f():
 | |
|             pass
 | |
|         # Since the function is local, lookup will fail
 | |
|         for proto in range(0, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             with self.assertRaises((AttributeError, pickle.PicklingError)):
 | |
|                 pickletools.dis(self.dumps(f, proto))
 | |
|         # Same without a __module__ attribute (exercises a different path
 | |
|         # in _pickle.c).
 | |
|         del f.__module__
 | |
|         for proto in range(0, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             with self.assertRaises((AttributeError, pickle.PicklingError)):
 | |
|                 pickletools.dis(self.dumps(f, proto))
 | |
|         # Yet a different path.
 | |
|         f.__name__ = f.__qualname__
 | |
|         for proto in range(0, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             with self.assertRaises((AttributeError, pickle.PicklingError)):
 | |
|                 pickletools.dis(self.dumps(f, proto))
 | |
| 
 | |
|     #
 | |
|     # PEP 574 tests below
 | |
|     #
 | |
| 
 | |
|     def buffer_like_objects(self):
 | |
|         # Yield buffer-like objects with the bytestring "abcdef" in them
 | |
|         bytestring = b"abcdefgh"
 | |
|         yield ZeroCopyBytes(bytestring)
 | |
|         yield ZeroCopyBytearray(bytestring)
 | |
|         if _testbuffer is not None:
 | |
|             items = list(bytestring)
 | |
|             value = int.from_bytes(bytestring, byteorder='little')
 | |
|             for flags in (0, _testbuffer.ND_WRITABLE):
 | |
|                 # 1-D, contiguous
 | |
|                 yield PicklableNDArray(items, format='B', shape=(8,),
 | |
|                                        flags=flags)
 | |
|                 # 2-D, C-contiguous
 | |
|                 yield PicklableNDArray(items, format='B', shape=(4, 2),
 | |
|                                        strides=(2, 1), flags=flags)
 | |
|                 # 2-D, Fortran-contiguous
 | |
|                 yield PicklableNDArray(items, format='B',
 | |
|                                        shape=(4, 2), strides=(1, 4),
 | |
|                                        flags=flags)
 | |
| 
 | |
|     def test_in_band_buffers(self):
 | |
|         # Test in-band buffers (PEP 574)
 | |
|         for obj in self.buffer_like_objects():
 | |
|             for proto in range(0, pickle.HIGHEST_PROTOCOL + 1):
 | |
|                 data = self.dumps(obj, proto)
 | |
|                 if obj.c_contiguous and proto >= 5:
 | |
|                     # The raw memory bytes are serialized in physical order
 | |
|                     self.assertIn(b"abcdefgh", data)
 | |
|                 self.assertEqual(count_opcode(pickle.NEXT_BUFFER, data), 0)
 | |
|                 if proto >= 5:
 | |
|                     self.assertEqual(count_opcode(pickle.SHORT_BINBYTES, data),
 | |
|                                      1 if obj.readonly else 0)
 | |
|                     self.assertEqual(count_opcode(pickle.BYTEARRAY8, data),
 | |
|                                      0 if obj.readonly else 1)
 | |
|                     # Return a true value from buffer_callback should have
 | |
|                     # the same effect
 | |
|                     def buffer_callback(obj):
 | |
|                         return True
 | |
|                     data2 = self.dumps(obj, proto,
 | |
|                                        buffer_callback=buffer_callback)
 | |
|                     self.assertEqual(data2, data)
 | |
| 
 | |
|                 new = self.loads(data)
 | |
|                 # It's a copy
 | |
|                 self.assertIsNot(new, obj)
 | |
|                 self.assertIs(type(new), type(obj))
 | |
|                 self.assertEqual(new, obj)
 | |
| 
 | |
|     # XXX Unfortunately cannot test non-contiguous array
 | |
|     # (see comment in PicklableNDArray.__reduce_ex__)
 | |
| 
 | |
|     def test_oob_buffers(self):
 | |
|         # Test out-of-band buffers (PEP 574)
 | |
|         for obj in self.buffer_like_objects():
 | |
|             for proto in range(0, 5):
 | |
|                 # Need protocol >= 5 for buffer_callback
 | |
|                 with self.assertRaises(ValueError):
 | |
|                     self.dumps(obj, proto,
 | |
|                                buffer_callback=[].append)
 | |
|             for proto in range(5, pickle.HIGHEST_PROTOCOL + 1):
 | |
|                 buffers = []
 | |
|                 buffer_callback = lambda pb: buffers.append(pb.raw())
 | |
|                 data = self.dumps(obj, proto,
 | |
|                                   buffer_callback=buffer_callback)
 | |
|                 self.assertNotIn(b"abcdefgh", data)
 | |
|                 self.assertEqual(count_opcode(pickle.SHORT_BINBYTES, data), 0)
 | |
|                 self.assertEqual(count_opcode(pickle.BYTEARRAY8, data), 0)
 | |
|                 self.assertEqual(count_opcode(pickle.NEXT_BUFFER, data), 1)
 | |
|                 self.assertEqual(count_opcode(pickle.READONLY_BUFFER, data),
 | |
|                                  1 if obj.readonly else 0)
 | |
| 
 | |
|                 if obj.c_contiguous:
 | |
|                     self.assertEqual(bytes(buffers[0]), b"abcdefgh")
 | |
|                 # Need buffers argument to unpickle properly
 | |
|                 with self.assertRaises(pickle.UnpicklingError):
 | |
|                     self.loads(data)
 | |
| 
 | |
|                 new = self.loads(data, buffers=buffers)
 | |
|                 if obj.zero_copy_reconstruct:
 | |
|                     # Zero-copy achieved
 | |
|                     self.assertIs(new, obj)
 | |
|                 else:
 | |
|                     self.assertIs(type(new), type(obj))
 | |
|                     self.assertEqual(new, obj)
 | |
|                 # Non-sequence buffers accepted too
 | |
|                 new = self.loads(data, buffers=iter(buffers))
 | |
|                 if obj.zero_copy_reconstruct:
 | |
|                     # Zero-copy achieved
 | |
|                     self.assertIs(new, obj)
 | |
|                 else:
 | |
|                     self.assertIs(type(new), type(obj))
 | |
|                     self.assertEqual(new, obj)
 | |
| 
 | |
|     def test_oob_buffers_writable_to_readonly(self):
 | |
|         # Test reconstructing readonly object from writable buffer
 | |
|         obj = ZeroCopyBytes(b"foobar")
 | |
|         for proto in range(5, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             buffers = []
 | |
|             buffer_callback = buffers.append
 | |
|             data = self.dumps(obj, proto, buffer_callback=buffer_callback)
 | |
| 
 | |
|             buffers = map(bytearray, buffers)
 | |
|             new = self.loads(data, buffers=buffers)
 | |
|             self.assertIs(type(new), type(obj))
 | |
|             self.assertEqual(new, obj)
 | |
| 
 | |
|     def test_picklebuffer_error(self):
 | |
|         # PickleBuffer forbidden with protocol < 5
 | |
|         pb = pickle.PickleBuffer(b"foobar")
 | |
|         for proto in range(0, 5):
 | |
|             with self.assertRaises(pickle.PickleError):
 | |
|                 self.dumps(pb, proto)
 | |
| 
 | |
|     def test_buffer_callback_error(self):
 | |
|         def buffer_callback(buffers):
 | |
|             1/0
 | |
|         pb = pickle.PickleBuffer(b"foobar")
 | |
|         with self.assertRaises(ZeroDivisionError):
 | |
|             self.dumps(pb, 5, buffer_callback=buffer_callback)
 | |
| 
 | |
|     def test_buffers_error(self):
 | |
|         pb = pickle.PickleBuffer(b"foobar")
 | |
|         for proto in range(5, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             data = self.dumps(pb, proto, buffer_callback=[].append)
 | |
|             # Non iterable buffers
 | |
|             with self.assertRaises(TypeError):
 | |
|                 self.loads(data, buffers=object())
 | |
|             # Buffer iterable exhausts too early
 | |
|             with self.assertRaises(pickle.UnpicklingError):
 | |
|                 self.loads(data, buffers=[])
 | |
| 
 | |
|     def test_inband_accept_default_buffers_argument(self):
 | |
|         for proto in range(5, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             data_pickled = self.dumps(1, proto, buffer_callback=None)
 | |
|             data = self.loads(data_pickled, buffers=None)
 | |
| 
 | |
|     @unittest.skipIf(np is None, "Test needs Numpy")
 | |
|     def test_buffers_numpy(self):
 | |
|         def check_no_copy(x, y):
 | |
|             np.testing.assert_equal(x, y)
 | |
|             self.assertEqual(x.ctypes.data, y.ctypes.data)
 | |
| 
 | |
|         def check_copy(x, y):
 | |
|             np.testing.assert_equal(x, y)
 | |
|             self.assertNotEqual(x.ctypes.data, y.ctypes.data)
 | |
| 
 | |
|         def check_array(arr):
 | |
|             # In-band
 | |
|             for proto in range(0, pickle.HIGHEST_PROTOCOL + 1):
 | |
|                 data = self.dumps(arr, proto)
 | |
|                 new = self.loads(data)
 | |
|                 check_copy(arr, new)
 | |
|             for proto in range(5, pickle.HIGHEST_PROTOCOL + 1):
 | |
|                 buffer_callback = lambda _: True
 | |
|                 data = self.dumps(arr, proto, buffer_callback=buffer_callback)
 | |
|                 new = self.loads(data)
 | |
|                 check_copy(arr, new)
 | |
|             # Out-of-band
 | |
|             for proto in range(5, pickle.HIGHEST_PROTOCOL + 1):
 | |
|                 buffers = []
 | |
|                 buffer_callback = buffers.append
 | |
|                 data = self.dumps(arr, proto, buffer_callback=buffer_callback)
 | |
|                 new = self.loads(data, buffers=buffers)
 | |
|                 if arr.flags.c_contiguous or arr.flags.f_contiguous:
 | |
|                     check_no_copy(arr, new)
 | |
|                 else:
 | |
|                     check_copy(arr, new)
 | |
| 
 | |
|         # 1-D
 | |
|         arr = np.arange(6)
 | |
|         check_array(arr)
 | |
|         # 1-D, non-contiguous
 | |
|         check_array(arr[::2])
 | |
|         # 2-D, C-contiguous
 | |
|         arr = np.arange(12).reshape((3, 4))
 | |
|         check_array(arr)
 | |
|         # 2-D, F-contiguous
 | |
|         check_array(arr.T)
 | |
|         # 2-D, non-contiguous
 | |
|         check_array(arr[::2])
 | |
| 
 | |
| 
 | |
| class BigmemPickleTests(unittest.TestCase):
 | |
| 
 | |
|     # Binary protocols can serialize longs of up to 2 GiB-1
 | |
| 
 | |
|     @bigmemtest(size=_2G, memuse=3.6, dry_run=False)
 | |
|     def test_huge_long_32b(self, size):
 | |
|         data = 1 << (8 * size)
 | |
|         try:
 | |
|             for proto in protocols:
 | |
|                 if proto < 2:
 | |
|                     continue
 | |
|                 with self.subTest(proto=proto):
 | |
|                     with self.assertRaises((ValueError, OverflowError)):
 | |
|                         self.dumps(data, protocol=proto)
 | |
|         finally:
 | |
|             data = None
 | |
| 
 | |
|     # Protocol 3 can serialize up to 4 GiB-1 as a bytes object
 | |
|     # (older protocols don't have a dedicated opcode for bytes and are
 | |
|     # too inefficient)
 | |
| 
 | |
|     @bigmemtest(size=_2G, memuse=2.5, dry_run=False)
 | |
|     def test_huge_bytes_32b(self, size):
 | |
|         data = b"abcd" * (size // 4)
 | |
|         try:
 | |
|             for proto in protocols:
 | |
|                 if proto < 3:
 | |
|                     continue
 | |
|                 with self.subTest(proto=proto):
 | |
|                     try:
 | |
|                         pickled = self.dumps(data, protocol=proto)
 | |
|                         header = (pickle.BINBYTES +
 | |
|                                   struct.pack("<I", len(data)))
 | |
|                         data_start = pickled.index(data)
 | |
|                         self.assertEqual(
 | |
|                             header,
 | |
|                             pickled[data_start-len(header):data_start])
 | |
|                     finally:
 | |
|                         pickled = None
 | |
|         finally:
 | |
|             data = None
 | |
| 
 | |
|     @bigmemtest(size=_4G, memuse=2.5, dry_run=False)
 | |
|     def test_huge_bytes_64b(self, size):
 | |
|         data = b"acbd" * (size // 4)
 | |
|         try:
 | |
|             for proto in protocols:
 | |
|                 if proto < 3:
 | |
|                     continue
 | |
|                 with self.subTest(proto=proto):
 | |
|                     if proto == 3:
 | |
|                         # Protocol 3 does not support large bytes objects.
 | |
|                         # Verify that we do not crash when processing one.
 | |
|                         with self.assertRaises((ValueError, OverflowError)):
 | |
|                             self.dumps(data, protocol=proto)
 | |
|                         continue
 | |
|                     try:
 | |
|                         pickled = self.dumps(data, protocol=proto)
 | |
|                         header = (pickle.BINBYTES8 +
 | |
|                                   struct.pack("<Q", len(data)))
 | |
|                         data_start = pickled.index(data)
 | |
|                         self.assertEqual(
 | |
|                             header,
 | |
|                             pickled[data_start-len(header):data_start])
 | |
|                     finally:
 | |
|                         pickled = None
 | |
|         finally:
 | |
|             data = None
 | |
| 
 | |
|     # All protocols use 1-byte per printable ASCII character; we add another
 | |
|     # byte because the encoded form has to be copied into the internal buffer.
 | |
| 
 | |
|     @bigmemtest(size=_2G, memuse=8, dry_run=False)
 | |
|     def test_huge_str_32b(self, size):
 | |
|         data = "abcd" * (size // 4)
 | |
|         try:
 | |
|             for proto in protocols:
 | |
|                 if proto == 0:
 | |
|                     continue
 | |
|                 with self.subTest(proto=proto):
 | |
|                     try:
 | |
|                         pickled = self.dumps(data, protocol=proto)
 | |
|                         header = (pickle.BINUNICODE +
 | |
|                                   struct.pack("<I", len(data)))
 | |
|                         data_start = pickled.index(b'abcd')
 | |
|                         self.assertEqual(
 | |
|                             header,
 | |
|                             pickled[data_start-len(header):data_start])
 | |
|                         self.assertEqual((pickled.rindex(b"abcd") + len(b"abcd") -
 | |
|                                           pickled.index(b"abcd")), len(data))
 | |
|                     finally:
 | |
|                         pickled = None
 | |
|         finally:
 | |
|             data = None
 | |
| 
 | |
|     # BINUNICODE (protocols 1, 2 and 3) cannot carry more than 2**32 - 1 bytes
 | |
|     # of utf-8 encoded unicode. BINUNICODE8 (protocol 4) supports these huge
 | |
|     # unicode strings however.
 | |
| 
 | |
|     @bigmemtest(size=_4G, memuse=8, dry_run=False)
 | |
|     def test_huge_str_64b(self, size):
 | |
|         data = "abcd" * (size // 4)
 | |
|         try:
 | |
|             for proto in protocols:
 | |
|                 if proto == 0:
 | |
|                     continue
 | |
|                 with self.subTest(proto=proto):
 | |
|                     if proto < 4:
 | |
|                         with self.assertRaises((ValueError, OverflowError)):
 | |
|                             self.dumps(data, protocol=proto)
 | |
|                         continue
 | |
|                     try:
 | |
|                         pickled = self.dumps(data, protocol=proto)
 | |
|                         header = (pickle.BINUNICODE8 +
 | |
|                                   struct.pack("<Q", len(data)))
 | |
|                         data_start = pickled.index(b'abcd')
 | |
|                         self.assertEqual(
 | |
|                             header,
 | |
|                             pickled[data_start-len(header):data_start])
 | |
|                         self.assertEqual((pickled.rindex(b"abcd") + len(b"abcd") -
 | |
|                                           pickled.index(b"abcd")), len(data))
 | |
|                     finally:
 | |
|                         pickled = None
 | |
|         finally:
 | |
|             data = None
 | |
| 
 | |
| 
 | |
| # Test classes for reduce_ex
 | |
| 
 | |
| class REX_one(object):
 | |
|     """No __reduce_ex__ here, but inheriting it from object"""
 | |
|     _reduce_called = 0
 | |
|     def __reduce__(self):
 | |
|         self._reduce_called = 1
 | |
|         return REX_one, ()
 | |
| 
 | |
| class REX_two(object):
 | |
|     """No __reduce__ here, but inheriting it from object"""
 | |
|     _proto = None
 | |
|     def __reduce_ex__(self, proto):
 | |
|         self._proto = proto
 | |
|         return REX_two, ()
 | |
| 
 | |
| class REX_three(object):
 | |
|     _proto = None
 | |
|     def __reduce_ex__(self, proto):
 | |
|         self._proto = proto
 | |
|         return REX_two, ()
 | |
|     def __reduce__(self):
 | |
|         raise TestFailed("This __reduce__ shouldn't be called")
 | |
| 
 | |
| class REX_four(object):
 | |
|     """Calling base class method should succeed"""
 | |
|     _proto = None
 | |
|     def __reduce_ex__(self, proto):
 | |
|         self._proto = proto
 | |
|         return object.__reduce_ex__(self, proto)
 | |
| 
 | |
| class REX_five(object):
 | |
|     """This one used to fail with infinite recursion"""
 | |
|     _reduce_called = 0
 | |
|     def __reduce__(self):
 | |
|         self._reduce_called = 1
 | |
|         return object.__reduce__(self)
 | |
| 
 | |
| class REX_six(object):
 | |
|     """This class is used to check the 4th argument (list iterator) of
 | |
|     the reduce protocol.
 | |
|     """
 | |
|     def __init__(self, items=None):
 | |
|         self.items = items if items is not None else []
 | |
|     def __eq__(self, other):
 | |
|         return type(self) is type(other) and self.items == other.items
 | |
|     def append(self, item):
 | |
|         self.items.append(item)
 | |
|     def __reduce__(self):
 | |
|         return type(self), (), None, iter(self.items), None
 | |
| 
 | |
| class REX_seven(object):
 | |
|     """This class is used to check the 5th argument (dict iterator) of
 | |
|     the reduce protocol.
 | |
|     """
 | |
|     def __init__(self, table=None):
 | |
|         self.table = table if table is not None else {}
 | |
|     def __eq__(self, other):
 | |
|         return type(self) is type(other) and self.table == other.table
 | |
|     def __setitem__(self, key, value):
 | |
|         self.table[key] = value
 | |
|     def __reduce__(self):
 | |
|         return type(self), (), None, None, iter(self.table.items())
 | |
| 
 | |
| class REX_state(object):
 | |
|     """This class is used to check the 3th argument (state) of
 | |
|     the reduce protocol.
 | |
|     """
 | |
|     def __init__(self, state=None):
 | |
|         self.state = state
 | |
|     def __eq__(self, other):
 | |
|         return type(self) is type(other) and self.state == other.state
 | |
|     def __setstate__(self, state):
 | |
|         self.state = state
 | |
|     def __reduce__(self):
 | |
|         return type(self), (), self.state
 | |
| 
 | |
| 
 | |
| # Test classes for newobj
 | |
| 
 | |
| class MyInt(int):
 | |
|     sample = 1
 | |
| 
 | |
| class MyFloat(float):
 | |
|     sample = 1.0
 | |
| 
 | |
| class MyComplex(complex):
 | |
|     sample = 1.0 + 0.0j
 | |
| 
 | |
| class MyStr(str):
 | |
|     sample = "hello"
 | |
| 
 | |
| class MyUnicode(str):
 | |
|     sample = "hello \u1234"
 | |
| 
 | |
| class MyTuple(tuple):
 | |
|     sample = (1, 2, 3)
 | |
| 
 | |
| class MyList(list):
 | |
|     sample = [1, 2, 3]
 | |
| 
 | |
| class MyDict(dict):
 | |
|     sample = {"a": 1, "b": 2}
 | |
| 
 | |
| class MySet(set):
 | |
|     sample = {"a", "b"}
 | |
| 
 | |
| class MyFrozenSet(frozenset):
 | |
|     sample = frozenset({"a", "b"})
 | |
| 
 | |
| myclasses = [MyInt, MyFloat,
 | |
|              MyComplex,
 | |
|              MyStr, MyUnicode,
 | |
|              MyTuple, MyList, MyDict, MySet, MyFrozenSet]
 | |
| 
 | |
| class MyIntWithNew(int):
 | |
|     def __new__(cls, value):
 | |
|         raise AssertionError
 | |
| 
 | |
| class MyIntWithNew2(MyIntWithNew):
 | |
|     __new__ = int.__new__
 | |
| 
 | |
| 
 | |
| class SlotList(MyList):
 | |
|     __slots__ = ["foo"]
 | |
| 
 | |
| class SimpleNewObj(int):
 | |
|     def __init__(self, *args, **kwargs):
 | |
|         # raise an error, to make sure this isn't called
 | |
|         raise TypeError("SimpleNewObj.__init__() didn't expect to get called")
 | |
|     def __eq__(self, other):
 | |
|         return int(self) == int(other) and self.__dict__ == other.__dict__
 | |
| 
 | |
| class ComplexNewObj(SimpleNewObj):
 | |
|     def __getnewargs__(self):
 | |
|         return ('%X' % self, 16)
 | |
| 
 | |
| class ComplexNewObjEx(SimpleNewObj):
 | |
|     def __getnewargs_ex__(self):
 | |
|         return ('%X' % self,), {'base': 16}
 | |
| 
 | |
| class BadGetattr:
 | |
|     def __getattr__(self, key):
 | |
|         self.foo
 | |
| 
 | |
| 
 | |
| class AbstractPickleModuleTests(unittest.TestCase):
 | |
| 
 | |
|     def test_dump_closed_file(self):
 | |
|         f = open(TESTFN, "wb")
 | |
|         try:
 | |
|             f.close()
 | |
|             self.assertRaises(ValueError, self.dump, 123, f)
 | |
|         finally:
 | |
|             support.unlink(TESTFN)
 | |
| 
 | |
|     def test_load_closed_file(self):
 | |
|         f = open(TESTFN, "wb")
 | |
|         try:
 | |
|             f.close()
 | |
|             self.assertRaises(ValueError, self.dump, 123, f)
 | |
|         finally:
 | |
|             support.unlink(TESTFN)
 | |
| 
 | |
|     def test_load_from_and_dump_to_file(self):
 | |
|         stream = io.BytesIO()
 | |
|         data = [123, {}, 124]
 | |
|         self.dump(data, stream)
 | |
|         stream.seek(0)
 | |
|         unpickled = self.load(stream)
 | |
|         self.assertEqual(unpickled, data)
 | |
| 
 | |
|     def test_highest_protocol(self):
 | |
|         # Of course this needs to be changed when HIGHEST_PROTOCOL changes.
 | |
|         self.assertEqual(pickle.HIGHEST_PROTOCOL, 5)
 | |
| 
 | |
|     def test_callapi(self):
 | |
|         f = io.BytesIO()
 | |
|         # With and without keyword arguments
 | |
|         self.dump(123, f, -1)
 | |
|         self.dump(123, file=f, protocol=-1)
 | |
|         self.dumps(123, -1)
 | |
|         self.dumps(123, protocol=-1)
 | |
|         self.Pickler(f, -1)
 | |
|         self.Pickler(f, protocol=-1)
 | |
| 
 | |
|     def test_dump_text_file(self):
 | |
|         f = open(TESTFN, "w")
 | |
|         try:
 | |
|             for proto in protocols:
 | |
|                 self.assertRaises(TypeError, self.dump, 123, f, proto)
 | |
|         finally:
 | |
|             f.close()
 | |
|             support.unlink(TESTFN)
 | |
| 
 | |
|     def test_incomplete_input(self):
 | |
|         s = io.BytesIO(b"X''.")
 | |
|         self.assertRaises((EOFError, struct.error, pickle.UnpicklingError), self.load, s)
 | |
| 
 | |
|     def test_bad_init(self):
 | |
|         # Test issue3664 (pickle can segfault from a badly initialized Pickler).
 | |
|         # Override initialization without calling __init__() of the superclass.
 | |
|         class BadPickler(self.Pickler):
 | |
|             def __init__(self): pass
 | |
| 
 | |
|         class BadUnpickler(self.Unpickler):
 | |
|             def __init__(self): pass
 | |
| 
 | |
|         self.assertRaises(pickle.PicklingError, BadPickler().dump, 0)
 | |
|         self.assertRaises(pickle.UnpicklingError, BadUnpickler().load)
 | |
| 
 | |
|     def check_dumps_loads_oob_buffers(self, dumps, loads):
 | |
|         # No need to do the full gamut of tests here, just enough to
 | |
|         # check that dumps() and loads() redirect their arguments
 | |
|         # to the underlying Pickler and Unpickler, respectively.
 | |
|         obj = ZeroCopyBytes(b"foo")
 | |
| 
 | |
|         for proto in range(0, 5):
 | |
|             # Need protocol >= 5 for buffer_callback
 | |
|             with self.assertRaises(ValueError):
 | |
|                 dumps(obj, protocol=proto,
 | |
|                       buffer_callback=[].append)
 | |
|         for proto in range(5, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             buffers = []
 | |
|             buffer_callback = buffers.append
 | |
|             data = dumps(obj, protocol=proto,
 | |
|                          buffer_callback=buffer_callback)
 | |
|             self.assertNotIn(b"foo", data)
 | |
|             self.assertEqual(bytes(buffers[0]), b"foo")
 | |
|             # Need buffers argument to unpickle properly
 | |
|             with self.assertRaises(pickle.UnpicklingError):
 | |
|                 loads(data)
 | |
|             new = loads(data, buffers=buffers)
 | |
|             self.assertIs(new, obj)
 | |
| 
 | |
|     def test_dumps_loads_oob_buffers(self):
 | |
|         # Test out-of-band buffers (PEP 574) with top-level dumps() and loads()
 | |
|         self.check_dumps_loads_oob_buffers(self.dumps, self.loads)
 | |
| 
 | |
|     def test_dump_load_oob_buffers(self):
 | |
|         # Test out-of-band buffers (PEP 574) with top-level dump() and load()
 | |
|         def dumps(obj, **kwargs):
 | |
|             f = io.BytesIO()
 | |
|             self.dump(obj, f, **kwargs)
 | |
|             return f.getvalue()
 | |
| 
 | |
|         def loads(data, **kwargs):
 | |
|             f = io.BytesIO(data)
 | |
|             return self.load(f, **kwargs)
 | |
| 
 | |
|         self.check_dumps_loads_oob_buffers(dumps, loads)
 | |
| 
 | |
| 
 | |
| class AbstractPersistentPicklerTests(unittest.TestCase):
 | |
| 
 | |
|     # This class defines persistent_id() and persistent_load()
 | |
|     # functions that should be used by the pickler.  All even integers
 | |
|     # are pickled using persistent ids.
 | |
| 
 | |
|     def persistent_id(self, object):
 | |
|         if isinstance(object, int) and object % 2 == 0:
 | |
|             self.id_count += 1
 | |
|             return str(object)
 | |
|         elif object == "test_false_value":
 | |
|             self.false_count += 1
 | |
|             return ""
 | |
|         else:
 | |
|             return None
 | |
| 
 | |
|     def persistent_load(self, oid):
 | |
|         if not oid:
 | |
|             self.load_false_count += 1
 | |
|             return "test_false_value"
 | |
|         else:
 | |
|             self.load_count += 1
 | |
|             object = int(oid)
 | |
|             assert object % 2 == 0
 | |
|             return object
 | |
| 
 | |
|     def test_persistence(self):
 | |
|         L = list(range(10)) + ["test_false_value"]
 | |
|         for proto in protocols:
 | |
|             self.id_count = 0
 | |
|             self.false_count = 0
 | |
|             self.load_false_count = 0
 | |
|             self.load_count = 0
 | |
|             self.assertEqual(self.loads(self.dumps(L, proto)), L)
 | |
|             self.assertEqual(self.id_count, 5)
 | |
|             self.assertEqual(self.false_count, 1)
 | |
|             self.assertEqual(self.load_count, 5)
 | |
|             self.assertEqual(self.load_false_count, 1)
 | |
| 
 | |
| 
 | |
| class AbstractIdentityPersistentPicklerTests(unittest.TestCase):
 | |
| 
 | |
|     def persistent_id(self, obj):
 | |
|         return obj
 | |
| 
 | |
|     def persistent_load(self, pid):
 | |
|         return pid
 | |
| 
 | |
|     def _check_return_correct_type(self, obj, proto):
 | |
|         unpickled = self.loads(self.dumps(obj, proto))
 | |
|         self.assertIsInstance(unpickled, type(obj))
 | |
|         self.assertEqual(unpickled, obj)
 | |
| 
 | |
|     def test_return_correct_type(self):
 | |
|         for proto in protocols:
 | |
|             # Protocol 0 supports only ASCII strings.
 | |
|             if proto == 0:
 | |
|                 self._check_return_correct_type("abc", 0)
 | |
|             else:
 | |
|                 for obj in [b"abc\n", "abc\n", -1, -1.1 * 0.1, str]:
 | |
|                     self._check_return_correct_type(obj, proto)
 | |
| 
 | |
|     def test_protocol0_is_ascii_only(self):
 | |
|         non_ascii_str = "\N{EMPTY SET}"
 | |
|         self.assertRaises(pickle.PicklingError, self.dumps, non_ascii_str, 0)
 | |
|         pickled = pickle.PERSID + non_ascii_str.encode('utf-8') + b'\n.'
 | |
|         self.assertRaises(pickle.UnpicklingError, self.loads, pickled)
 | |
| 
 | |
| 
 | |
| class AbstractPicklerUnpicklerObjectTests(unittest.TestCase):
 | |
| 
 | |
|     pickler_class = None
 | |
|     unpickler_class = None
 | |
| 
 | |
|     def setUp(self):
 | |
|         assert self.pickler_class
 | |
|         assert self.unpickler_class
 | |
| 
 | |
|     def test_clear_pickler_memo(self):
 | |
|         # To test whether clear_memo() has any effect, we pickle an object,
 | |
|         # then pickle it again without clearing the memo; the two serialized
 | |
|         # forms should be different. If we clear_memo() and then pickle the
 | |
|         # object again, the third serialized form should be identical to the
 | |
|         # first one we obtained.
 | |
|         data = ["abcdefg", "abcdefg", 44]
 | |
|         for proto in protocols:
 | |
|             f = io.BytesIO()
 | |
|             pickler = self.pickler_class(f, proto)
 | |
| 
 | |
|             pickler.dump(data)
 | |
|             first_pickled = f.getvalue()
 | |
| 
 | |
|             # Reset BytesIO object.
 | |
|             f.seek(0)
 | |
|             f.truncate()
 | |
| 
 | |
|             pickler.dump(data)
 | |
|             second_pickled = f.getvalue()
 | |
| 
 | |
|             # Reset the Pickler and BytesIO objects.
 | |
|             pickler.clear_memo()
 | |
|             f.seek(0)
 | |
|             f.truncate()
 | |
| 
 | |
|             pickler.dump(data)
 | |
|             third_pickled = f.getvalue()
 | |
| 
 | |
|             self.assertNotEqual(first_pickled, second_pickled)
 | |
|             self.assertEqual(first_pickled, third_pickled)
 | |
| 
 | |
|     def test_priming_pickler_memo(self):
 | |
|         # Verify that we can set the Pickler's memo attribute.
 | |
|         data = ["abcdefg", "abcdefg", 44]
 | |
|         f = io.BytesIO()
 | |
|         pickler = self.pickler_class(f)
 | |
| 
 | |
|         pickler.dump(data)
 | |
|         first_pickled = f.getvalue()
 | |
| 
 | |
|         f = io.BytesIO()
 | |
|         primed = self.pickler_class(f)
 | |
|         primed.memo = pickler.memo
 | |
| 
 | |
|         primed.dump(data)
 | |
|         primed_pickled = f.getvalue()
 | |
| 
 | |
|         self.assertNotEqual(first_pickled, primed_pickled)
 | |
| 
 | |
|     def test_priming_unpickler_memo(self):
 | |
|         # Verify that we can set the Unpickler's memo attribute.
 | |
|         data = ["abcdefg", "abcdefg", 44]
 | |
|         f = io.BytesIO()
 | |
|         pickler = self.pickler_class(f)
 | |
| 
 | |
|         pickler.dump(data)
 | |
|         first_pickled = f.getvalue()
 | |
| 
 | |
|         f = io.BytesIO()
 | |
|         primed = self.pickler_class(f)
 | |
|         primed.memo = pickler.memo
 | |
| 
 | |
|         primed.dump(data)
 | |
|         primed_pickled = f.getvalue()
 | |
| 
 | |
|         unpickler = self.unpickler_class(io.BytesIO(first_pickled))
 | |
|         unpickled_data1 = unpickler.load()
 | |
| 
 | |
|         self.assertEqual(unpickled_data1, data)
 | |
| 
 | |
|         primed = self.unpickler_class(io.BytesIO(primed_pickled))
 | |
|         primed.memo = unpickler.memo
 | |
|         unpickled_data2 = primed.load()
 | |
| 
 | |
|         primed.memo.clear()
 | |
| 
 | |
|         self.assertEqual(unpickled_data2, data)
 | |
|         self.assertTrue(unpickled_data2 is unpickled_data1)
 | |
| 
 | |
|     def test_reusing_unpickler_objects(self):
 | |
|         data1 = ["abcdefg", "abcdefg", 44]
 | |
|         f = io.BytesIO()
 | |
|         pickler = self.pickler_class(f)
 | |
|         pickler.dump(data1)
 | |
|         pickled1 = f.getvalue()
 | |
| 
 | |
|         data2 = ["abcdefg", 44, 44]
 | |
|         f = io.BytesIO()
 | |
|         pickler = self.pickler_class(f)
 | |
|         pickler.dump(data2)
 | |
|         pickled2 = f.getvalue()
 | |
| 
 | |
|         f = io.BytesIO()
 | |
|         f.write(pickled1)
 | |
|         f.seek(0)
 | |
|         unpickler = self.unpickler_class(f)
 | |
|         self.assertEqual(unpickler.load(), data1)
 | |
| 
 | |
|         f.seek(0)
 | |
|         f.truncate()
 | |
|         f.write(pickled2)
 | |
|         f.seek(0)
 | |
|         self.assertEqual(unpickler.load(), data2)
 | |
| 
 | |
|     def _check_multiple_unpicklings(self, ioclass, *, seekable=True):
 | |
|         for proto in protocols:
 | |
|             with self.subTest(proto=proto):
 | |
|                 data1 = [(x, str(x)) for x in range(2000)] + [b"abcde", len]
 | |
|                 f = ioclass()
 | |
|                 pickler = self.pickler_class(f, protocol=proto)
 | |
|                 pickler.dump(data1)
 | |
|                 pickled = f.getvalue()
 | |
| 
 | |
|                 N = 5
 | |
|                 f = ioclass(pickled * N)
 | |
|                 unpickler = self.unpickler_class(f)
 | |
|                 for i in range(N):
 | |
|                     if seekable:
 | |
|                         pos = f.tell()
 | |
|                     self.assertEqual(unpickler.load(), data1)
 | |
|                     if seekable:
 | |
|                         self.assertEqual(f.tell(), pos + len(pickled))
 | |
|                 self.assertRaises(EOFError, unpickler.load)
 | |
| 
 | |
|     def test_multiple_unpicklings_seekable(self):
 | |
|         self._check_multiple_unpicklings(io.BytesIO)
 | |
| 
 | |
|     def test_multiple_unpicklings_unseekable(self):
 | |
|         self._check_multiple_unpicklings(UnseekableIO, seekable=False)
 | |
| 
 | |
|     def test_multiple_unpicklings_minimal(self):
 | |
|         # File-like object that doesn't support peek() and readinto()
 | |
|         # (bpo-39681)
 | |
|         self._check_multiple_unpicklings(MinimalIO, seekable=False)
 | |
| 
 | |
|     def test_unpickling_buffering_readline(self):
 | |
|         # Issue #12687: the unpickler's buffering logic could fail with
 | |
|         # text mode opcodes.
 | |
|         data = list(range(10))
 | |
|         for proto in protocols:
 | |
|             for buf_size in range(1, 11):
 | |
|                 f = io.BufferedRandom(io.BytesIO(), buffer_size=buf_size)
 | |
|                 pickler = self.pickler_class(f, protocol=proto)
 | |
|                 pickler.dump(data)
 | |
|                 f.seek(0)
 | |
|                 unpickler = self.unpickler_class(f)
 | |
|                 self.assertEqual(unpickler.load(), data)
 | |
| 
 | |
| 
 | |
| # Tests for dispatch_table attribute
 | |
| 
 | |
| REDUCE_A = 'reduce_A'
 | |
| 
 | |
| class AAA(object):
 | |
|     def __reduce__(self):
 | |
|         return str, (REDUCE_A,)
 | |
| 
 | |
| class BBB(object):
 | |
|     def __init__(self):
 | |
|         # Add an instance attribute to enable state-saving routines at pickling
 | |
|         # time.
 | |
|         self.a = "some attribute"
 | |
| 
 | |
|     def __setstate__(self, state):
 | |
|         self.a = "BBB.__setstate__"
 | |
| 
 | |
| 
 | |
| def setstate_bbb(obj, state):
 | |
|     """Custom state setter for BBB objects
 | |
| 
 | |
|     Such callable may be created by other persons than the ones who created the
 | |
|     BBB class. If passed as the state_setter item of a custom reducer, this
 | |
|     allows for custom state setting behavior of BBB objects. One can think of
 | |
|     it as the analogous of list_setitems or dict_setitems but for foreign
 | |
|     classes/functions.
 | |
|     """
 | |
|     obj.a = "custom state_setter"
 | |
| 
 | |
| 
 | |
| 
 | |
| class AbstractCustomPicklerClass:
 | |
|     """Pickler implementing a reducing hook using reducer_override."""
 | |
|     def reducer_override(self, obj):
 | |
|         obj_name = getattr(obj, "__name__", None)
 | |
| 
 | |
|         if obj_name == 'f':
 | |
|             # asking the pickler to save f as 5
 | |
|             return int, (5, )
 | |
| 
 | |
|         if obj_name == 'MyClass':
 | |
|             return str, ('some str',)
 | |
| 
 | |
|         elif obj_name == 'g':
 | |
|             # in this case, the callback returns an invalid result (not a 2-5
 | |
|             # tuple or a string), the pickler should raise a proper error.
 | |
|             return False
 | |
| 
 | |
|         elif obj_name == 'h':
 | |
|             # Simulate a case when the reducer fails. The error should
 | |
|             # be propagated to the original ``dump`` call.
 | |
|             raise ValueError('The reducer just failed')
 | |
| 
 | |
|         return NotImplemented
 | |
| 
 | |
| class AbstractHookTests(unittest.TestCase):
 | |
|     def test_pickler_hook(self):
 | |
|         # test the ability of a custom, user-defined CPickler subclass to
 | |
|         # override the default reducing routines of any type using the method
 | |
|         # reducer_override
 | |
| 
 | |
|         def f():
 | |
|             pass
 | |
| 
 | |
|         def g():
 | |
|             pass
 | |
| 
 | |
|         def h():
 | |
|             pass
 | |
| 
 | |
|         class MyClass:
 | |
|             pass
 | |
| 
 | |
|         for proto in range(0, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             with self.subTest(proto=proto):
 | |
|                 bio = io.BytesIO()
 | |
|                 p = self.pickler_class(bio, proto)
 | |
| 
 | |
|                 p.dump([f, MyClass, math.log])
 | |
|                 new_f, some_str, math_log = pickle.loads(bio.getvalue())
 | |
| 
 | |
|                 self.assertEqual(new_f, 5)
 | |
|                 self.assertEqual(some_str, 'some str')
 | |
|                 # math.log does not have its usual reducer overriden, so the
 | |
|                 # custom reduction callback should silently direct the pickler
 | |
|                 # to the default pickling by attribute, by returning
 | |
|                 # NotImplemented
 | |
|                 self.assertIs(math_log, math.log)
 | |
| 
 | |
|                 with self.assertRaises(pickle.PicklingError):
 | |
|                     p.dump(g)
 | |
| 
 | |
|                 with self.assertRaisesRegex(
 | |
|                         ValueError, 'The reducer just failed'):
 | |
|                     p.dump(h)
 | |
| 
 | |
|     @support.cpython_only
 | |
|     def test_reducer_override_no_reference_cycle(self):
 | |
|         # bpo-39492: reducer_override used to induce a spurious reference cycle
 | |
|         # inside the Pickler object, that could prevent all serialized objects
 | |
|         # from being garbage-collected without explicity invoking gc.collect.
 | |
| 
 | |
|         for proto in range(0, pickle.HIGHEST_PROTOCOL + 1):
 | |
|             with self.subTest(proto=proto):
 | |
|                 def f():
 | |
|                     pass
 | |
| 
 | |
|                 wr = weakref.ref(f)
 | |
| 
 | |
|                 bio = io.BytesIO()
 | |
|                 p = self.pickler_class(bio, proto)
 | |
|                 p.dump(f)
 | |
|                 new_f = pickle.loads(bio.getvalue())
 | |
|                 assert new_f == 5
 | |
| 
 | |
|                 del p
 | |
|                 del f
 | |
| 
 | |
|                 self.assertIsNone(wr())
 | |
| 
 | |
| 
 | |
| class AbstractDispatchTableTests(unittest.TestCase):
 | |
| 
 | |
|     def test_default_dispatch_table(self):
 | |
|         # No dispatch_table attribute by default
 | |
|         f = io.BytesIO()
 | |
|         p = self.pickler_class(f, 0)
 | |
|         with self.assertRaises(AttributeError):
 | |
|             p.dispatch_table
 | |
|         self.assertFalse(hasattr(p, 'dispatch_table'))
 | |
| 
 | |
|     def test_class_dispatch_table(self):
 | |
|         # A dispatch_table attribute can be specified class-wide
 | |
|         dt = self.get_dispatch_table()
 | |
| 
 | |
|         class MyPickler(self.pickler_class):
 | |
|             dispatch_table = dt
 | |
| 
 | |
|         def dumps(obj, protocol=None):
 | |
|             f = io.BytesIO()
 | |
|             p = MyPickler(f, protocol)
 | |
|             self.assertEqual(p.dispatch_table, dt)
 | |
|             p.dump(obj)
 | |
|             return f.getvalue()
 | |
| 
 | |
|         self._test_dispatch_table(dumps, dt)
 | |
| 
 | |
|     def test_instance_dispatch_table(self):
 | |
|         # A dispatch_table attribute can also be specified instance-wide
 | |
|         dt = self.get_dispatch_table()
 | |
| 
 | |
|         def dumps(obj, protocol=None):
 | |
|             f = io.BytesIO()
 | |
|             p = self.pickler_class(f, protocol)
 | |
|             p.dispatch_table = dt
 | |
|             self.assertEqual(p.dispatch_table, dt)
 | |
|             p.dump(obj)
 | |
|             return f.getvalue()
 | |
| 
 | |
|         self._test_dispatch_table(dumps, dt)
 | |
| 
 | |
|     def _test_dispatch_table(self, dumps, dispatch_table):
 | |
|         def custom_load_dump(obj):
 | |
|             return pickle.loads(dumps(obj, 0))
 | |
| 
 | |
|         def default_load_dump(obj):
 | |
|             return pickle.loads(pickle.dumps(obj, 0))
 | |
| 
 | |
|         # pickling complex numbers using protocol 0 relies on copyreg
 | |
|         # so check pickling a complex number still works
 | |
|         z = 1 + 2j
 | |
|         self.assertEqual(custom_load_dump(z), z)
 | |
|         self.assertEqual(default_load_dump(z), z)
 | |
| 
 | |
|         # modify pickling of complex
 | |
|         REDUCE_1 = 'reduce_1'
 | |
|         def reduce_1(obj):
 | |
|             return str, (REDUCE_1,)
 | |
|         dispatch_table[complex] = reduce_1
 | |
|         self.assertEqual(custom_load_dump(z), REDUCE_1)
 | |
|         self.assertEqual(default_load_dump(z), z)
 | |
| 
 | |
|         # check picklability of AAA and BBB
 | |
|         a = AAA()
 | |
|         b = BBB()
 | |
|         self.assertEqual(custom_load_dump(a), REDUCE_A)
 | |
|         self.assertIsInstance(custom_load_dump(b), BBB)
 | |
|         self.assertEqual(default_load_dump(a), REDUCE_A)
 | |
|         self.assertIsInstance(default_load_dump(b), BBB)
 | |
| 
 | |
|         # modify pickling of BBB
 | |
|         dispatch_table[BBB] = reduce_1
 | |
|         self.assertEqual(custom_load_dump(a), REDUCE_A)
 | |
|         self.assertEqual(custom_load_dump(b), REDUCE_1)
 | |
|         self.assertEqual(default_load_dump(a), REDUCE_A)
 | |
|         self.assertIsInstance(default_load_dump(b), BBB)
 | |
| 
 | |
|         # revert pickling of BBB and modify pickling of AAA
 | |
|         REDUCE_2 = 'reduce_2'
 | |
|         def reduce_2(obj):
 | |
|             return str, (REDUCE_2,)
 | |
|         dispatch_table[AAA] = reduce_2
 | |
|         del dispatch_table[BBB]
 | |
|         self.assertEqual(custom_load_dump(a), REDUCE_2)
 | |
|         self.assertIsInstance(custom_load_dump(b), BBB)
 | |
|         self.assertEqual(default_load_dump(a), REDUCE_A)
 | |
|         self.assertIsInstance(default_load_dump(b), BBB)
 | |
| 
 | |
|         # End-to-end testing of save_reduce with the state_setter keyword
 | |
|         # argument. This is a dispatch_table test as the primary goal of
 | |
|         # state_setter is to tweak objects reduction behavior.
 | |
|         # In particular, state_setter is useful when the default __setstate__
 | |
|         # behavior is not flexible enough.
 | |
| 
 | |
|         # No custom reducer for b has been registered for now, so
 | |
|         # BBB.__setstate__ should be used at unpickling time
 | |
|         self.assertEqual(default_load_dump(b).a, "BBB.__setstate__")
 | |
| 
 | |
|         def reduce_bbb(obj):
 | |
|             return BBB, (), obj.__dict__, None, None, setstate_bbb
 | |
| 
 | |
|         dispatch_table[BBB] = reduce_bbb
 | |
| 
 | |
|         # The custom reducer reduce_bbb includes a state setter, that should
 | |
|         # have priority over BBB.__setstate__
 | |
|         self.assertEqual(custom_load_dump(b).a, "custom state_setter")
 | |
| 
 | |
| 
 | |
| if __name__ == "__main__":
 | |
|     # Print some stuff that can be used to rewrite DATA{0,1,2}
 | |
|     from pickletools import dis
 | |
|     x = create_data()
 | |
|     for i in range(pickle.HIGHEST_PROTOCOL+1):
 | |
|         p = pickle.dumps(x, i)
 | |
|         print("DATA{0} = (".format(i))
 | |
|         for j in range(0, len(p), 20):
 | |
|             b = bytes(p[j:j+20])
 | |
|             print("    {0!r}".format(b))
 | |
|         print(")")
 | |
|         print()
 | |
|         print("# Disassembly of DATA{0}".format(i))
 | |
|         print("DATA{0}_DIS = \"\"\"\\".format(i))
 | |
|         dis(p)
 | |
|         print("\"\"\"")
 | |
|         print()
 |