8015 lines
229 KiB
C
8015 lines
229 KiB
C
/* pickle accelerator C extensor: _pickle module.
|
|
*
|
|
* It is built as a built-in module (Py_BUILD_CORE_BUILTIN define) on Windows
|
|
* and as an extension module (Py_BUILD_CORE_MODULE define) on other
|
|
* platforms. */
|
|
|
|
#if !defined(Py_BUILD_CORE_BUILTIN) && !defined(Py_BUILD_CORE_MODULE)
|
|
# error "Py_BUILD_CORE_BUILTIN or Py_BUILD_CORE_MODULE must be defined"
|
|
#endif
|
|
|
|
#include "Python.h"
|
|
#include "structmember.h" // PyMemberDef
|
|
|
|
PyDoc_STRVAR(pickle_module_doc,
|
|
"Optimized C implementation for the Python pickle module.");
|
|
|
|
/*[clinic input]
|
|
module _pickle
|
|
class _pickle.Pickler "PicklerObject *" "&Pickler_Type"
|
|
class _pickle.PicklerMemoProxy "PicklerMemoProxyObject *" "&PicklerMemoProxyType"
|
|
class _pickle.Unpickler "UnpicklerObject *" "&Unpickler_Type"
|
|
class _pickle.UnpicklerMemoProxy "UnpicklerMemoProxyObject *" "&UnpicklerMemoProxyType"
|
|
[clinic start generated code]*/
|
|
/*[clinic end generated code: output=da39a3ee5e6b4b0d input=4b3e113468a58e6c]*/
|
|
|
|
/* Bump HIGHEST_PROTOCOL when new opcodes are added to the pickle protocol.
|
|
Bump DEFAULT_PROTOCOL only when the oldest still supported version of Python
|
|
already includes it. */
|
|
enum {
|
|
HIGHEST_PROTOCOL = 5,
|
|
DEFAULT_PROTOCOL = 4
|
|
};
|
|
|
|
/* Pickle opcodes. These must be kept updated with pickle.py.
|
|
Extensive docs are in pickletools.py. */
|
|
enum opcode {
|
|
MARK = '(',
|
|
STOP = '.',
|
|
POP = '0',
|
|
POP_MARK = '1',
|
|
DUP = '2',
|
|
FLOAT = 'F',
|
|
INT = 'I',
|
|
BININT = 'J',
|
|
BININT1 = 'K',
|
|
LONG = 'L',
|
|
BININT2 = 'M',
|
|
NONE = 'N',
|
|
PERSID = 'P',
|
|
BINPERSID = 'Q',
|
|
REDUCE = 'R',
|
|
STRING = 'S',
|
|
BINSTRING = 'T',
|
|
SHORT_BINSTRING = 'U',
|
|
UNICODE = 'V',
|
|
BINUNICODE = 'X',
|
|
APPEND = 'a',
|
|
BUILD = 'b',
|
|
GLOBAL = 'c',
|
|
DICT = 'd',
|
|
EMPTY_DICT = '}',
|
|
APPENDS = 'e',
|
|
GET = 'g',
|
|
BINGET = 'h',
|
|
INST = 'i',
|
|
LONG_BINGET = 'j',
|
|
LIST = 'l',
|
|
EMPTY_LIST = ']',
|
|
OBJ = 'o',
|
|
PUT = 'p',
|
|
BINPUT = 'q',
|
|
LONG_BINPUT = 'r',
|
|
SETITEM = 's',
|
|
TUPLE = 't',
|
|
EMPTY_TUPLE = ')',
|
|
SETITEMS = 'u',
|
|
BINFLOAT = 'G',
|
|
|
|
/* Protocol 2. */
|
|
PROTO = '\x80',
|
|
NEWOBJ = '\x81',
|
|
EXT1 = '\x82',
|
|
EXT2 = '\x83',
|
|
EXT4 = '\x84',
|
|
TUPLE1 = '\x85',
|
|
TUPLE2 = '\x86',
|
|
TUPLE3 = '\x87',
|
|
NEWTRUE = '\x88',
|
|
NEWFALSE = '\x89',
|
|
LONG1 = '\x8a',
|
|
LONG4 = '\x8b',
|
|
|
|
/* Protocol 3 (Python 3.x) */
|
|
BINBYTES = 'B',
|
|
SHORT_BINBYTES = 'C',
|
|
|
|
/* Protocol 4 */
|
|
SHORT_BINUNICODE = '\x8c',
|
|
BINUNICODE8 = '\x8d',
|
|
BINBYTES8 = '\x8e',
|
|
EMPTY_SET = '\x8f',
|
|
ADDITEMS = '\x90',
|
|
FROZENSET = '\x91',
|
|
NEWOBJ_EX = '\x92',
|
|
STACK_GLOBAL = '\x93',
|
|
MEMOIZE = '\x94',
|
|
FRAME = '\x95',
|
|
|
|
/* Protocol 5 */
|
|
BYTEARRAY8 = '\x96',
|
|
NEXT_BUFFER = '\x97',
|
|
READONLY_BUFFER = '\x98'
|
|
};
|
|
|
|
enum {
|
|
/* Keep in synch with pickle.Pickler._BATCHSIZE. This is how many elements
|
|
batch_list/dict() pumps out before doing APPENDS/SETITEMS. Nothing will
|
|
break if this gets out of synch with pickle.py, but it's unclear that would
|
|
help anything either. */
|
|
BATCHSIZE = 1000,
|
|
|
|
/* Nesting limit until Pickler, when running in "fast mode", starts
|
|
checking for self-referential data-structures. */
|
|
FAST_NESTING_LIMIT = 50,
|
|
|
|
/* Initial size of the write buffer of Pickler. */
|
|
WRITE_BUF_SIZE = 4096,
|
|
|
|
/* Prefetch size when unpickling (disabled on unpeekable streams) */
|
|
PREFETCH = 8192 * 16,
|
|
|
|
FRAME_SIZE_MIN = 4,
|
|
FRAME_SIZE_TARGET = 64 * 1024,
|
|
FRAME_HEADER_SIZE = 9
|
|
};
|
|
|
|
/*************************************************************************/
|
|
|
|
/* State of the pickle module, per PEP 3121. */
|
|
typedef struct {
|
|
/* Exception classes for pickle. */
|
|
PyObject *PickleError;
|
|
PyObject *PicklingError;
|
|
PyObject *UnpicklingError;
|
|
|
|
/* copyreg.dispatch_table, {type_object: pickling_function} */
|
|
PyObject *dispatch_table;
|
|
|
|
/* For the extension opcodes EXT1, EXT2 and EXT4. */
|
|
|
|
/* copyreg._extension_registry, {(module_name, function_name): code} */
|
|
PyObject *extension_registry;
|
|
/* copyreg._extension_cache, {code: object} */
|
|
PyObject *extension_cache;
|
|
/* copyreg._inverted_registry, {code: (module_name, function_name)} */
|
|
PyObject *inverted_registry;
|
|
|
|
/* Import mappings for compatibility with Python 2.x */
|
|
|
|
/* _compat_pickle.NAME_MAPPING,
|
|
{(oldmodule, oldname): (newmodule, newname)} */
|
|
PyObject *name_mapping_2to3;
|
|
/* _compat_pickle.IMPORT_MAPPING, {oldmodule: newmodule} */
|
|
PyObject *import_mapping_2to3;
|
|
/* Same, but with REVERSE_NAME_MAPPING / REVERSE_IMPORT_MAPPING */
|
|
PyObject *name_mapping_3to2;
|
|
PyObject *import_mapping_3to2;
|
|
|
|
/* codecs.encode, used for saving bytes in older protocols */
|
|
PyObject *codecs_encode;
|
|
/* builtins.getattr, used for saving nested names with protocol < 4 */
|
|
PyObject *getattr;
|
|
/* functools.partial, used for implementing __newobj_ex__ with protocols
|
|
2 and 3 */
|
|
PyObject *partial;
|
|
} PickleState;
|
|
|
|
/* Forward declaration of the _pickle module definition. */
|
|
static struct PyModuleDef _picklemodule;
|
|
|
|
/* Given a module object, get its per-module state. */
|
|
static PickleState *
|
|
_Pickle_GetState(PyObject *module)
|
|
{
|
|
return (PickleState *)PyModule_GetState(module);
|
|
}
|
|
|
|
/* Find the module instance imported in the currently running sub-interpreter
|
|
and get its state. */
|
|
static PickleState *
|
|
_Pickle_GetGlobalState(void)
|
|
{
|
|
return _Pickle_GetState(PyState_FindModule(&_picklemodule));
|
|
}
|
|
|
|
/* Clear the given pickle module state. */
|
|
static void
|
|
_Pickle_ClearState(PickleState *st)
|
|
{
|
|
Py_CLEAR(st->PickleError);
|
|
Py_CLEAR(st->PicklingError);
|
|
Py_CLEAR(st->UnpicklingError);
|
|
Py_CLEAR(st->dispatch_table);
|
|
Py_CLEAR(st->extension_registry);
|
|
Py_CLEAR(st->extension_cache);
|
|
Py_CLEAR(st->inverted_registry);
|
|
Py_CLEAR(st->name_mapping_2to3);
|
|
Py_CLEAR(st->import_mapping_2to3);
|
|
Py_CLEAR(st->name_mapping_3to2);
|
|
Py_CLEAR(st->import_mapping_3to2);
|
|
Py_CLEAR(st->codecs_encode);
|
|
Py_CLEAR(st->getattr);
|
|
Py_CLEAR(st->partial);
|
|
}
|
|
|
|
/* Initialize the given pickle module state. */
|
|
static int
|
|
_Pickle_InitState(PickleState *st)
|
|
{
|
|
PyObject *copyreg = NULL;
|
|
PyObject *compat_pickle = NULL;
|
|
PyObject *codecs = NULL;
|
|
PyObject *functools = NULL;
|
|
_Py_IDENTIFIER(getattr);
|
|
|
|
st->getattr = _PyEval_GetBuiltinId(&PyId_getattr);
|
|
if (st->getattr == NULL)
|
|
goto error;
|
|
|
|
copyreg = PyImport_ImportModule("copyreg");
|
|
if (!copyreg)
|
|
goto error;
|
|
st->dispatch_table = PyObject_GetAttrString(copyreg, "dispatch_table");
|
|
if (!st->dispatch_table)
|
|
goto error;
|
|
if (!PyDict_CheckExact(st->dispatch_table)) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"copyreg.dispatch_table should be a dict, not %.200s",
|
|
Py_TYPE(st->dispatch_table)->tp_name);
|
|
goto error;
|
|
}
|
|
st->extension_registry = \
|
|
PyObject_GetAttrString(copyreg, "_extension_registry");
|
|
if (!st->extension_registry)
|
|
goto error;
|
|
if (!PyDict_CheckExact(st->extension_registry)) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"copyreg._extension_registry should be a dict, "
|
|
"not %.200s", Py_TYPE(st->extension_registry)->tp_name);
|
|
goto error;
|
|
}
|
|
st->inverted_registry = \
|
|
PyObject_GetAttrString(copyreg, "_inverted_registry");
|
|
if (!st->inverted_registry)
|
|
goto error;
|
|
if (!PyDict_CheckExact(st->inverted_registry)) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"copyreg._inverted_registry should be a dict, "
|
|
"not %.200s", Py_TYPE(st->inverted_registry)->tp_name);
|
|
goto error;
|
|
}
|
|
st->extension_cache = PyObject_GetAttrString(copyreg, "_extension_cache");
|
|
if (!st->extension_cache)
|
|
goto error;
|
|
if (!PyDict_CheckExact(st->extension_cache)) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"copyreg._extension_cache should be a dict, "
|
|
"not %.200s", Py_TYPE(st->extension_cache)->tp_name);
|
|
goto error;
|
|
}
|
|
Py_CLEAR(copyreg);
|
|
|
|
/* Load the 2.x -> 3.x stdlib module mapping tables */
|
|
compat_pickle = PyImport_ImportModule("_compat_pickle");
|
|
if (!compat_pickle)
|
|
goto error;
|
|
st->name_mapping_2to3 = \
|
|
PyObject_GetAttrString(compat_pickle, "NAME_MAPPING");
|
|
if (!st->name_mapping_2to3)
|
|
goto error;
|
|
if (!PyDict_CheckExact(st->name_mapping_2to3)) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"_compat_pickle.NAME_MAPPING should be a dict, not %.200s",
|
|
Py_TYPE(st->name_mapping_2to3)->tp_name);
|
|
goto error;
|
|
}
|
|
st->import_mapping_2to3 = \
|
|
PyObject_GetAttrString(compat_pickle, "IMPORT_MAPPING");
|
|
if (!st->import_mapping_2to3)
|
|
goto error;
|
|
if (!PyDict_CheckExact(st->import_mapping_2to3)) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"_compat_pickle.IMPORT_MAPPING should be a dict, "
|
|
"not %.200s", Py_TYPE(st->import_mapping_2to3)->tp_name);
|
|
goto error;
|
|
}
|
|
/* ... and the 3.x -> 2.x mapping tables */
|
|
st->name_mapping_3to2 = \
|
|
PyObject_GetAttrString(compat_pickle, "REVERSE_NAME_MAPPING");
|
|
if (!st->name_mapping_3to2)
|
|
goto error;
|
|
if (!PyDict_CheckExact(st->name_mapping_3to2)) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"_compat_pickle.REVERSE_NAME_MAPPING should be a dict, "
|
|
"not %.200s", Py_TYPE(st->name_mapping_3to2)->tp_name);
|
|
goto error;
|
|
}
|
|
st->import_mapping_3to2 = \
|
|
PyObject_GetAttrString(compat_pickle, "REVERSE_IMPORT_MAPPING");
|
|
if (!st->import_mapping_3to2)
|
|
goto error;
|
|
if (!PyDict_CheckExact(st->import_mapping_3to2)) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"_compat_pickle.REVERSE_IMPORT_MAPPING should be a dict, "
|
|
"not %.200s", Py_TYPE(st->import_mapping_3to2)->tp_name);
|
|
goto error;
|
|
}
|
|
Py_CLEAR(compat_pickle);
|
|
|
|
codecs = PyImport_ImportModule("codecs");
|
|
if (codecs == NULL)
|
|
goto error;
|
|
st->codecs_encode = PyObject_GetAttrString(codecs, "encode");
|
|
if (st->codecs_encode == NULL) {
|
|
goto error;
|
|
}
|
|
if (!PyCallable_Check(st->codecs_encode)) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"codecs.encode should be a callable, not %.200s",
|
|
Py_TYPE(st->codecs_encode)->tp_name);
|
|
goto error;
|
|
}
|
|
Py_CLEAR(codecs);
|
|
|
|
functools = PyImport_ImportModule("functools");
|
|
if (!functools)
|
|
goto error;
|
|
st->partial = PyObject_GetAttrString(functools, "partial");
|
|
if (!st->partial)
|
|
goto error;
|
|
Py_CLEAR(functools);
|
|
|
|
return 0;
|
|
|
|
error:
|
|
Py_CLEAR(copyreg);
|
|
Py_CLEAR(compat_pickle);
|
|
Py_CLEAR(codecs);
|
|
Py_CLEAR(functools);
|
|
_Pickle_ClearState(st);
|
|
return -1;
|
|
}
|
|
|
|
/* Helper for calling a function with a single argument quickly.
|
|
|
|
This function steals the reference of the given argument. */
|
|
static PyObject *
|
|
_Pickle_FastCall(PyObject *func, PyObject *obj)
|
|
{
|
|
PyObject *result;
|
|
|
|
result = PyObject_CallOneArg(func, obj);
|
|
Py_DECREF(obj);
|
|
return result;
|
|
}
|
|
|
|
/*************************************************************************/
|
|
|
|
/* Retrieve and deconstruct a method for avoiding a reference cycle
|
|
(pickler -> bound method of pickler -> pickler) */
|
|
static int
|
|
init_method_ref(PyObject *self, _Py_Identifier *name,
|
|
PyObject **method_func, PyObject **method_self)
|
|
{
|
|
PyObject *func, *func2;
|
|
int ret;
|
|
|
|
/* *method_func and *method_self should be consistent. All refcount decrements
|
|
should be occurred after setting *method_self and *method_func. */
|
|
ret = _PyObject_LookupAttrId(self, name, &func);
|
|
if (func == NULL) {
|
|
*method_self = NULL;
|
|
Py_CLEAR(*method_func);
|
|
return ret;
|
|
}
|
|
|
|
if (PyMethod_Check(func) && PyMethod_GET_SELF(func) == self) {
|
|
/* Deconstruct a bound Python method */
|
|
func2 = PyMethod_GET_FUNCTION(func);
|
|
Py_INCREF(func2);
|
|
*method_self = self; /* borrowed */
|
|
Py_XSETREF(*method_func, func2);
|
|
Py_DECREF(func);
|
|
return 0;
|
|
}
|
|
else {
|
|
*method_self = NULL;
|
|
Py_XSETREF(*method_func, func);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
/* Bind a method if it was deconstructed */
|
|
static PyObject *
|
|
reconstruct_method(PyObject *func, PyObject *self)
|
|
{
|
|
if (self) {
|
|
return PyMethod_New(func, self);
|
|
}
|
|
else {
|
|
Py_INCREF(func);
|
|
return func;
|
|
}
|
|
}
|
|
|
|
static PyObject *
|
|
call_method(PyObject *func, PyObject *self, PyObject *obj)
|
|
{
|
|
if (self) {
|
|
return PyObject_CallFunctionObjArgs(func, self, obj, NULL);
|
|
}
|
|
else {
|
|
return PyObject_CallOneArg(func, obj);
|
|
}
|
|
}
|
|
|
|
/*************************************************************************/
|
|
|
|
/* Internal data type used as the unpickling stack. */
|
|
typedef struct {
|
|
PyObject_VAR_HEAD
|
|
PyObject **data;
|
|
int mark_set; /* is MARK set? */
|
|
Py_ssize_t fence; /* position of top MARK or 0 */
|
|
Py_ssize_t allocated; /* number of slots in data allocated */
|
|
} Pdata;
|
|
|
|
static void
|
|
Pdata_dealloc(Pdata *self)
|
|
{
|
|
Py_ssize_t i = Py_SIZE(self);
|
|
while (--i >= 0) {
|
|
Py_DECREF(self->data[i]);
|
|
}
|
|
PyMem_FREE(self->data);
|
|
PyObject_Del(self);
|
|
}
|
|
|
|
static PyTypeObject Pdata_Type = {
|
|
PyVarObject_HEAD_INIT(NULL, 0)
|
|
"_pickle.Pdata", /*tp_name*/
|
|
sizeof(Pdata), /*tp_basicsize*/
|
|
sizeof(PyObject *), /*tp_itemsize*/
|
|
(destructor)Pdata_dealloc, /*tp_dealloc*/
|
|
};
|
|
|
|
static PyObject *
|
|
Pdata_New(void)
|
|
{
|
|
Pdata *self;
|
|
|
|
if (!(self = PyObject_New(Pdata, &Pdata_Type)))
|
|
return NULL;
|
|
Py_SET_SIZE(self, 0);
|
|
self->mark_set = 0;
|
|
self->fence = 0;
|
|
self->allocated = 8;
|
|
self->data = PyMem_MALLOC(self->allocated * sizeof(PyObject *));
|
|
if (self->data)
|
|
return (PyObject *)self;
|
|
Py_DECREF(self);
|
|
return PyErr_NoMemory();
|
|
}
|
|
|
|
|
|
/* Retain only the initial clearto items. If clearto >= the current
|
|
* number of items, this is a (non-erroneous) NOP.
|
|
*/
|
|
static int
|
|
Pdata_clear(Pdata *self, Py_ssize_t clearto)
|
|
{
|
|
Py_ssize_t i = Py_SIZE(self);
|
|
|
|
assert(clearto >= self->fence);
|
|
if (clearto >= i)
|
|
return 0;
|
|
|
|
while (--i >= clearto) {
|
|
Py_CLEAR(self->data[i]);
|
|
}
|
|
Py_SET_SIZE(self, clearto);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
Pdata_grow(Pdata *self)
|
|
{
|
|
PyObject **data = self->data;
|
|
size_t allocated = (size_t)self->allocated;
|
|
size_t new_allocated;
|
|
|
|
new_allocated = (allocated >> 3) + 6;
|
|
/* check for integer overflow */
|
|
if (new_allocated > (size_t)PY_SSIZE_T_MAX - allocated)
|
|
goto nomemory;
|
|
new_allocated += allocated;
|
|
PyMem_RESIZE(data, PyObject *, new_allocated);
|
|
if (data == NULL)
|
|
goto nomemory;
|
|
|
|
self->data = data;
|
|
self->allocated = (Py_ssize_t)new_allocated;
|
|
return 0;
|
|
|
|
nomemory:
|
|
PyErr_NoMemory();
|
|
return -1;
|
|
}
|
|
|
|
static int
|
|
Pdata_stack_underflow(Pdata *self)
|
|
{
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->UnpicklingError,
|
|
self->mark_set ?
|
|
"unexpected MARK found" :
|
|
"unpickling stack underflow");
|
|
return -1;
|
|
}
|
|
|
|
/* D is a Pdata*. Pop the topmost element and store it into V, which
|
|
* must be an lvalue holding PyObject*. On stack underflow, UnpicklingError
|
|
* is raised and V is set to NULL.
|
|
*/
|
|
static PyObject *
|
|
Pdata_pop(Pdata *self)
|
|
{
|
|
if (Py_SIZE(self) <= self->fence) {
|
|
Pdata_stack_underflow(self);
|
|
return NULL;
|
|
}
|
|
Py_SET_SIZE(self, Py_SIZE(self) - 1);
|
|
return self->data[Py_SIZE(self)];
|
|
}
|
|
#define PDATA_POP(D, V) do { (V) = Pdata_pop((D)); } while (0)
|
|
|
|
static int
|
|
Pdata_push(Pdata *self, PyObject *obj)
|
|
{
|
|
if (Py_SIZE(self) == self->allocated && Pdata_grow(self) < 0) {
|
|
return -1;
|
|
}
|
|
self->data[Py_SIZE(self)] = obj;
|
|
Py_SET_SIZE(self, Py_SIZE(self) + 1);
|
|
return 0;
|
|
}
|
|
|
|
/* Push an object on stack, transferring its ownership to the stack. */
|
|
#define PDATA_PUSH(D, O, ER) do { \
|
|
if (Pdata_push((D), (O)) < 0) return (ER); } while(0)
|
|
|
|
/* Push an object on stack, adding a new reference to the object. */
|
|
#define PDATA_APPEND(D, O, ER) do { \
|
|
Py_INCREF((O)); \
|
|
if (Pdata_push((D), (O)) < 0) return (ER); } while(0)
|
|
|
|
static PyObject *
|
|
Pdata_poptuple(Pdata *self, Py_ssize_t start)
|
|
{
|
|
PyObject *tuple;
|
|
Py_ssize_t len, i, j;
|
|
|
|
if (start < self->fence) {
|
|
Pdata_stack_underflow(self);
|
|
return NULL;
|
|
}
|
|
len = Py_SIZE(self) - start;
|
|
tuple = PyTuple_New(len);
|
|
if (tuple == NULL)
|
|
return NULL;
|
|
for (i = start, j = 0; j < len; i++, j++)
|
|
PyTuple_SET_ITEM(tuple, j, self->data[i]);
|
|
|
|
Py_SET_SIZE(self, start);
|
|
return tuple;
|
|
}
|
|
|
|
static PyObject *
|
|
Pdata_poplist(Pdata *self, Py_ssize_t start)
|
|
{
|
|
PyObject *list;
|
|
Py_ssize_t len, i, j;
|
|
|
|
len = Py_SIZE(self) - start;
|
|
list = PyList_New(len);
|
|
if (list == NULL)
|
|
return NULL;
|
|
for (i = start, j = 0; j < len; i++, j++)
|
|
PyList_SET_ITEM(list, j, self->data[i]);
|
|
|
|
Py_SET_SIZE(self, start);
|
|
return list;
|
|
}
|
|
|
|
typedef struct {
|
|
PyObject *me_key;
|
|
Py_ssize_t me_value;
|
|
} PyMemoEntry;
|
|
|
|
typedef struct {
|
|
size_t mt_mask;
|
|
size_t mt_used;
|
|
size_t mt_allocated;
|
|
PyMemoEntry *mt_table;
|
|
} PyMemoTable;
|
|
|
|
typedef struct PicklerObject {
|
|
PyObject_HEAD
|
|
PyMemoTable *memo; /* Memo table, keep track of the seen
|
|
objects to support self-referential objects
|
|
pickling. */
|
|
PyObject *pers_func; /* persistent_id() method, can be NULL */
|
|
PyObject *pers_func_self; /* borrowed reference to self if pers_func
|
|
is an unbound method, NULL otherwise */
|
|
PyObject *dispatch_table; /* private dispatch_table, can be NULL */
|
|
PyObject *reducer_override; /* hook for invoking user-defined callbacks
|
|
instead of save_global when pickling
|
|
functions and classes*/
|
|
|
|
PyObject *write; /* write() method of the output stream. */
|
|
PyObject *output_buffer; /* Write into a local bytearray buffer before
|
|
flushing to the stream. */
|
|
Py_ssize_t output_len; /* Length of output_buffer. */
|
|
Py_ssize_t max_output_len; /* Allocation size of output_buffer. */
|
|
int proto; /* Pickle protocol number, >= 0 */
|
|
int bin; /* Boolean, true if proto > 0 */
|
|
int framing; /* True when framing is enabled, proto >= 4 */
|
|
Py_ssize_t frame_start; /* Position in output_buffer where the
|
|
current frame begins. -1 if there
|
|
is no frame currently open. */
|
|
|
|
Py_ssize_t buf_size; /* Size of the current buffered pickle data */
|
|
int fast; /* Enable fast mode if set to a true value.
|
|
The fast mode disable the usage of memo,
|
|
therefore speeding the pickling process by
|
|
not generating superfluous PUT opcodes. It
|
|
should not be used if with self-referential
|
|
objects. */
|
|
int fast_nesting;
|
|
int fix_imports; /* Indicate whether Pickler should fix
|
|
the name of globals for Python 2.x. */
|
|
PyObject *fast_memo;
|
|
PyObject *buffer_callback; /* Callback for out-of-band buffers, or NULL */
|
|
} PicklerObject;
|
|
|
|
typedef struct UnpicklerObject {
|
|
PyObject_HEAD
|
|
Pdata *stack; /* Pickle data stack, store unpickled objects. */
|
|
|
|
/* The unpickler memo is just an array of PyObject *s. Using a dict
|
|
is unnecessary, since the keys are contiguous ints. */
|
|
PyObject **memo;
|
|
size_t memo_size; /* Capacity of the memo array */
|
|
size_t memo_len; /* Number of objects in the memo */
|
|
|
|
PyObject *pers_func; /* persistent_load() method, can be NULL. */
|
|
PyObject *pers_func_self; /* borrowed reference to self if pers_func
|
|
is an unbound method, NULL otherwise */
|
|
|
|
Py_buffer buffer;
|
|
char *input_buffer;
|
|
char *input_line;
|
|
Py_ssize_t input_len;
|
|
Py_ssize_t next_read_idx;
|
|
Py_ssize_t prefetched_idx; /* index of first prefetched byte */
|
|
|
|
PyObject *read; /* read() method of the input stream. */
|
|
PyObject *readinto; /* readinto() method of the input stream. */
|
|
PyObject *readline; /* readline() method of the input stream. */
|
|
PyObject *peek; /* peek() method of the input stream, or NULL */
|
|
PyObject *buffers; /* iterable of out-of-band buffers, or NULL */
|
|
|
|
char *encoding; /* Name of the encoding to be used for
|
|
decoding strings pickled using Python
|
|
2.x. The default value is "ASCII" */
|
|
char *errors; /* Name of errors handling scheme to used when
|
|
decoding strings. The default value is
|
|
"strict". */
|
|
Py_ssize_t *marks; /* Mark stack, used for unpickling container
|
|
objects. */
|
|
Py_ssize_t num_marks; /* Number of marks in the mark stack. */
|
|
Py_ssize_t marks_size; /* Current allocated size of the mark stack. */
|
|
int proto; /* Protocol of the pickle loaded. */
|
|
int fix_imports; /* Indicate whether Unpickler should fix
|
|
the name of globals pickled by Python 2.x. */
|
|
} UnpicklerObject;
|
|
|
|
typedef struct {
|
|
PyObject_HEAD
|
|
PicklerObject *pickler; /* Pickler whose memo table we're proxying. */
|
|
} PicklerMemoProxyObject;
|
|
|
|
typedef struct {
|
|
PyObject_HEAD
|
|
UnpicklerObject *unpickler;
|
|
} UnpicklerMemoProxyObject;
|
|
|
|
/* Forward declarations */
|
|
static int save(PicklerObject *, PyObject *, int);
|
|
static int save_reduce(PicklerObject *, PyObject *, PyObject *);
|
|
static PyTypeObject Pickler_Type;
|
|
static PyTypeObject Unpickler_Type;
|
|
|
|
#include "clinic/_pickle.c.h"
|
|
|
|
/*************************************************************************
|
|
A custom hashtable mapping void* to Python ints. This is used by the pickler
|
|
for memoization. Using a custom hashtable rather than PyDict allows us to skip
|
|
a bunch of unnecessary object creation. This makes a huge performance
|
|
difference. */
|
|
|
|
#define MT_MINSIZE 8
|
|
#define PERTURB_SHIFT 5
|
|
|
|
|
|
static PyMemoTable *
|
|
PyMemoTable_New(void)
|
|
{
|
|
PyMemoTable *memo = PyMem_MALLOC(sizeof(PyMemoTable));
|
|
if (memo == NULL) {
|
|
PyErr_NoMemory();
|
|
return NULL;
|
|
}
|
|
|
|
memo->mt_used = 0;
|
|
memo->mt_allocated = MT_MINSIZE;
|
|
memo->mt_mask = MT_MINSIZE - 1;
|
|
memo->mt_table = PyMem_MALLOC(MT_MINSIZE * sizeof(PyMemoEntry));
|
|
if (memo->mt_table == NULL) {
|
|
PyMem_FREE(memo);
|
|
PyErr_NoMemory();
|
|
return NULL;
|
|
}
|
|
memset(memo->mt_table, 0, MT_MINSIZE * sizeof(PyMemoEntry));
|
|
|
|
return memo;
|
|
}
|
|
|
|
static PyMemoTable *
|
|
PyMemoTable_Copy(PyMemoTable *self)
|
|
{
|
|
PyMemoTable *new = PyMemoTable_New();
|
|
if (new == NULL)
|
|
return NULL;
|
|
|
|
new->mt_used = self->mt_used;
|
|
new->mt_allocated = self->mt_allocated;
|
|
new->mt_mask = self->mt_mask;
|
|
/* The table we get from _New() is probably smaller than we wanted.
|
|
Free it and allocate one that's the right size. */
|
|
PyMem_FREE(new->mt_table);
|
|
new->mt_table = PyMem_NEW(PyMemoEntry, self->mt_allocated);
|
|
if (new->mt_table == NULL) {
|
|
PyMem_FREE(new);
|
|
PyErr_NoMemory();
|
|
return NULL;
|
|
}
|
|
for (size_t i = 0; i < self->mt_allocated; i++) {
|
|
Py_XINCREF(self->mt_table[i].me_key);
|
|
}
|
|
memcpy(new->mt_table, self->mt_table,
|
|
sizeof(PyMemoEntry) * self->mt_allocated);
|
|
|
|
return new;
|
|
}
|
|
|
|
static Py_ssize_t
|
|
PyMemoTable_Size(PyMemoTable *self)
|
|
{
|
|
return self->mt_used;
|
|
}
|
|
|
|
static int
|
|
PyMemoTable_Clear(PyMemoTable *self)
|
|
{
|
|
Py_ssize_t i = self->mt_allocated;
|
|
|
|
while (--i >= 0) {
|
|
Py_XDECREF(self->mt_table[i].me_key);
|
|
}
|
|
self->mt_used = 0;
|
|
memset(self->mt_table, 0, self->mt_allocated * sizeof(PyMemoEntry));
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
PyMemoTable_Del(PyMemoTable *self)
|
|
{
|
|
if (self == NULL)
|
|
return;
|
|
PyMemoTable_Clear(self);
|
|
|
|
PyMem_FREE(self->mt_table);
|
|
PyMem_FREE(self);
|
|
}
|
|
|
|
/* Since entries cannot be deleted from this hashtable, _PyMemoTable_Lookup()
|
|
can be considerably simpler than dictobject.c's lookdict(). */
|
|
static PyMemoEntry *
|
|
_PyMemoTable_Lookup(PyMemoTable *self, PyObject *key)
|
|
{
|
|
size_t i;
|
|
size_t perturb;
|
|
size_t mask = self->mt_mask;
|
|
PyMemoEntry *table = self->mt_table;
|
|
PyMemoEntry *entry;
|
|
Py_hash_t hash = (Py_hash_t)key >> 3;
|
|
|
|
i = hash & mask;
|
|
entry = &table[i];
|
|
if (entry->me_key == NULL || entry->me_key == key)
|
|
return entry;
|
|
|
|
for (perturb = hash; ; perturb >>= PERTURB_SHIFT) {
|
|
i = (i << 2) + i + perturb + 1;
|
|
entry = &table[i & mask];
|
|
if (entry->me_key == NULL || entry->me_key == key)
|
|
return entry;
|
|
}
|
|
Py_UNREACHABLE();
|
|
}
|
|
|
|
/* Returns -1 on failure, 0 on success. */
|
|
static int
|
|
_PyMemoTable_ResizeTable(PyMemoTable *self, size_t min_size)
|
|
{
|
|
PyMemoEntry *oldtable = NULL;
|
|
PyMemoEntry *oldentry, *newentry;
|
|
size_t new_size = MT_MINSIZE;
|
|
size_t to_process;
|
|
|
|
assert(min_size > 0);
|
|
|
|
if (min_size > PY_SSIZE_T_MAX) {
|
|
PyErr_NoMemory();
|
|
return -1;
|
|
}
|
|
|
|
/* Find the smallest valid table size >= min_size. */
|
|
while (new_size < min_size) {
|
|
new_size <<= 1;
|
|
}
|
|
/* new_size needs to be a power of two. */
|
|
assert((new_size & (new_size - 1)) == 0);
|
|
|
|
/* Allocate new table. */
|
|
oldtable = self->mt_table;
|
|
self->mt_table = PyMem_NEW(PyMemoEntry, new_size);
|
|
if (self->mt_table == NULL) {
|
|
self->mt_table = oldtable;
|
|
PyErr_NoMemory();
|
|
return -1;
|
|
}
|
|
self->mt_allocated = new_size;
|
|
self->mt_mask = new_size - 1;
|
|
memset(self->mt_table, 0, sizeof(PyMemoEntry) * new_size);
|
|
|
|
/* Copy entries from the old table. */
|
|
to_process = self->mt_used;
|
|
for (oldentry = oldtable; to_process > 0; oldentry++) {
|
|
if (oldentry->me_key != NULL) {
|
|
to_process--;
|
|
/* newentry is a pointer to a chunk of the new
|
|
mt_table, so we're setting the key:value pair
|
|
in-place. */
|
|
newentry = _PyMemoTable_Lookup(self, oldentry->me_key);
|
|
newentry->me_key = oldentry->me_key;
|
|
newentry->me_value = oldentry->me_value;
|
|
}
|
|
}
|
|
|
|
/* Deallocate the old table. */
|
|
PyMem_FREE(oldtable);
|
|
return 0;
|
|
}
|
|
|
|
/* Returns NULL on failure, a pointer to the value otherwise. */
|
|
static Py_ssize_t *
|
|
PyMemoTable_Get(PyMemoTable *self, PyObject *key)
|
|
{
|
|
PyMemoEntry *entry = _PyMemoTable_Lookup(self, key);
|
|
if (entry->me_key == NULL)
|
|
return NULL;
|
|
return &entry->me_value;
|
|
}
|
|
|
|
/* Returns -1 on failure, 0 on success. */
|
|
static int
|
|
PyMemoTable_Set(PyMemoTable *self, PyObject *key, Py_ssize_t value)
|
|
{
|
|
PyMemoEntry *entry;
|
|
|
|
assert(key != NULL);
|
|
|
|
entry = _PyMemoTable_Lookup(self, key);
|
|
if (entry->me_key != NULL) {
|
|
entry->me_value = value;
|
|
return 0;
|
|
}
|
|
Py_INCREF(key);
|
|
entry->me_key = key;
|
|
entry->me_value = value;
|
|
self->mt_used++;
|
|
|
|
/* If we added a key, we can safely resize. Otherwise just return!
|
|
* If used >= 2/3 size, adjust size. Normally, this quaduples the size.
|
|
*
|
|
* Quadrupling the size improves average table sparseness
|
|
* (reducing collisions) at the cost of some memory. It also halves
|
|
* the number of expensive resize operations in a growing memo table.
|
|
*
|
|
* Very large memo tables (over 50K items) use doubling instead.
|
|
* This may help applications with severe memory constraints.
|
|
*/
|
|
if (SIZE_MAX / 3 >= self->mt_used && self->mt_used * 3 < self->mt_allocated * 2) {
|
|
return 0;
|
|
}
|
|
// self->mt_used is always < PY_SSIZE_T_MAX, so this can't overflow.
|
|
size_t desired_size = (self->mt_used > 50000 ? 2 : 4) * self->mt_used;
|
|
return _PyMemoTable_ResizeTable(self, desired_size);
|
|
}
|
|
|
|
#undef MT_MINSIZE
|
|
#undef PERTURB_SHIFT
|
|
|
|
/*************************************************************************/
|
|
|
|
|
|
static int
|
|
_Pickler_ClearBuffer(PicklerObject *self)
|
|
{
|
|
Py_XSETREF(self->output_buffer,
|
|
PyBytes_FromStringAndSize(NULL, self->max_output_len));
|
|
if (self->output_buffer == NULL)
|
|
return -1;
|
|
self->output_len = 0;
|
|
self->frame_start = -1;
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
_write_size64(char *out, size_t value)
|
|
{
|
|
size_t i;
|
|
|
|
Py_BUILD_ASSERT(sizeof(size_t) <= 8);
|
|
|
|
for (i = 0; i < sizeof(size_t); i++) {
|
|
out[i] = (unsigned char)((value >> (8 * i)) & 0xff);
|
|
}
|
|
for (i = sizeof(size_t); i < 8; i++) {
|
|
out[i] = 0;
|
|
}
|
|
}
|
|
|
|
static int
|
|
_Pickler_CommitFrame(PicklerObject *self)
|
|
{
|
|
size_t frame_len;
|
|
char *qdata;
|
|
|
|
if (!self->framing || self->frame_start == -1)
|
|
return 0;
|
|
frame_len = self->output_len - self->frame_start - FRAME_HEADER_SIZE;
|
|
qdata = PyBytes_AS_STRING(self->output_buffer) + self->frame_start;
|
|
if (frame_len >= FRAME_SIZE_MIN) {
|
|
qdata[0] = FRAME;
|
|
_write_size64(qdata + 1, frame_len);
|
|
}
|
|
else {
|
|
memmove(qdata, qdata + FRAME_HEADER_SIZE, frame_len);
|
|
self->output_len -= FRAME_HEADER_SIZE;
|
|
}
|
|
self->frame_start = -1;
|
|
return 0;
|
|
}
|
|
|
|
static PyObject *
|
|
_Pickler_GetString(PicklerObject *self)
|
|
{
|
|
PyObject *output_buffer = self->output_buffer;
|
|
|
|
assert(self->output_buffer != NULL);
|
|
|
|
if (_Pickler_CommitFrame(self))
|
|
return NULL;
|
|
|
|
self->output_buffer = NULL;
|
|
/* Resize down to exact size */
|
|
if (_PyBytes_Resize(&output_buffer, self->output_len) < 0)
|
|
return NULL;
|
|
return output_buffer;
|
|
}
|
|
|
|
static int
|
|
_Pickler_FlushToFile(PicklerObject *self)
|
|
{
|
|
PyObject *output, *result;
|
|
|
|
assert(self->write != NULL);
|
|
|
|
/* This will commit the frame first */
|
|
output = _Pickler_GetString(self);
|
|
if (output == NULL)
|
|
return -1;
|
|
|
|
result = _Pickle_FastCall(self->write, output);
|
|
Py_XDECREF(result);
|
|
return (result == NULL) ? -1 : 0;
|
|
}
|
|
|
|
static int
|
|
_Pickler_OpcodeBoundary(PicklerObject *self)
|
|
{
|
|
Py_ssize_t frame_len;
|
|
|
|
if (!self->framing || self->frame_start == -1) {
|
|
return 0;
|
|
}
|
|
frame_len = self->output_len - self->frame_start - FRAME_HEADER_SIZE;
|
|
if (frame_len >= FRAME_SIZE_TARGET) {
|
|
if(_Pickler_CommitFrame(self)) {
|
|
return -1;
|
|
}
|
|
/* Flush the content of the committed frame to the underlying
|
|
* file and reuse the pickler buffer for the next frame so as
|
|
* to limit memory usage when dumping large complex objects to
|
|
* a file.
|
|
*
|
|
* self->write is NULL when called via dumps.
|
|
*/
|
|
if (self->write != NULL) {
|
|
if (_Pickler_FlushToFile(self) < 0) {
|
|
return -1;
|
|
}
|
|
if (_Pickler_ClearBuffer(self) < 0) {
|
|
return -1;
|
|
}
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static Py_ssize_t
|
|
_Pickler_Write(PicklerObject *self, const char *s, Py_ssize_t data_len)
|
|
{
|
|
Py_ssize_t i, n, required;
|
|
char *buffer;
|
|
int need_new_frame;
|
|
|
|
assert(s != NULL);
|
|
need_new_frame = (self->framing && self->frame_start == -1);
|
|
|
|
if (need_new_frame)
|
|
n = data_len + FRAME_HEADER_SIZE;
|
|
else
|
|
n = data_len;
|
|
|
|
required = self->output_len + n;
|
|
if (required > self->max_output_len) {
|
|
/* Make place in buffer for the pickle chunk */
|
|
if (self->output_len >= PY_SSIZE_T_MAX / 2 - n) {
|
|
PyErr_NoMemory();
|
|
return -1;
|
|
}
|
|
self->max_output_len = (self->output_len + n) / 2 * 3;
|
|
if (_PyBytes_Resize(&self->output_buffer, self->max_output_len) < 0)
|
|
return -1;
|
|
}
|
|
buffer = PyBytes_AS_STRING(self->output_buffer);
|
|
if (need_new_frame) {
|
|
/* Setup new frame */
|
|
Py_ssize_t frame_start = self->output_len;
|
|
self->frame_start = frame_start;
|
|
for (i = 0; i < FRAME_HEADER_SIZE; i++) {
|
|
/* Write an invalid value, for debugging */
|
|
buffer[frame_start + i] = 0xFE;
|
|
}
|
|
self->output_len += FRAME_HEADER_SIZE;
|
|
}
|
|
if (data_len < 8) {
|
|
/* This is faster than memcpy when the string is short. */
|
|
for (i = 0; i < data_len; i++) {
|
|
buffer[self->output_len + i] = s[i];
|
|
}
|
|
}
|
|
else {
|
|
memcpy(buffer + self->output_len, s, data_len);
|
|
}
|
|
self->output_len += data_len;
|
|
return data_len;
|
|
}
|
|
|
|
static PicklerObject *
|
|
_Pickler_New(void)
|
|
{
|
|
PicklerObject *self;
|
|
|
|
self = PyObject_GC_New(PicklerObject, &Pickler_Type);
|
|
if (self == NULL)
|
|
return NULL;
|
|
|
|
self->pers_func = NULL;
|
|
self->dispatch_table = NULL;
|
|
self->buffer_callback = NULL;
|
|
self->write = NULL;
|
|
self->proto = 0;
|
|
self->bin = 0;
|
|
self->framing = 0;
|
|
self->frame_start = -1;
|
|
self->fast = 0;
|
|
self->fast_nesting = 0;
|
|
self->fix_imports = 0;
|
|
self->fast_memo = NULL;
|
|
self->max_output_len = WRITE_BUF_SIZE;
|
|
self->output_len = 0;
|
|
self->reducer_override = NULL;
|
|
|
|
self->memo = PyMemoTable_New();
|
|
self->output_buffer = PyBytes_FromStringAndSize(NULL,
|
|
self->max_output_len);
|
|
|
|
if (self->memo == NULL || self->output_buffer == NULL) {
|
|
Py_DECREF(self);
|
|
return NULL;
|
|
}
|
|
|
|
PyObject_GC_Track(self);
|
|
return self;
|
|
}
|
|
|
|
static int
|
|
_Pickler_SetProtocol(PicklerObject *self, PyObject *protocol, int fix_imports)
|
|
{
|
|
long proto;
|
|
|
|
if (protocol == Py_None) {
|
|
proto = DEFAULT_PROTOCOL;
|
|
}
|
|
else {
|
|
proto = PyLong_AsLong(protocol);
|
|
if (proto < 0) {
|
|
if (proto == -1 && PyErr_Occurred())
|
|
return -1;
|
|
proto = HIGHEST_PROTOCOL;
|
|
}
|
|
else if (proto > HIGHEST_PROTOCOL) {
|
|
PyErr_Format(PyExc_ValueError, "pickle protocol must be <= %d",
|
|
HIGHEST_PROTOCOL);
|
|
return -1;
|
|
}
|
|
}
|
|
self->proto = (int)proto;
|
|
self->bin = proto > 0;
|
|
self->fix_imports = fix_imports && proto < 3;
|
|
return 0;
|
|
}
|
|
|
|
/* Returns -1 (with an exception set) on failure, 0 on success. This may
|
|
be called once on a freshly created Pickler. */
|
|
static int
|
|
_Pickler_SetOutputStream(PicklerObject *self, PyObject *file)
|
|
{
|
|
_Py_IDENTIFIER(write);
|
|
assert(file != NULL);
|
|
if (_PyObject_LookupAttrId(file, &PyId_write, &self->write) < 0) {
|
|
return -1;
|
|
}
|
|
if (self->write == NULL) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"file must have a 'write' attribute");
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
_Pickler_SetBufferCallback(PicklerObject *self, PyObject *buffer_callback)
|
|
{
|
|
if (buffer_callback == Py_None) {
|
|
buffer_callback = NULL;
|
|
}
|
|
if (buffer_callback != NULL && self->proto < 5) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"buffer_callback needs protocol >= 5");
|
|
return -1;
|
|
}
|
|
|
|
Py_XINCREF(buffer_callback);
|
|
self->buffer_callback = buffer_callback;
|
|
return 0;
|
|
}
|
|
|
|
/* Returns the size of the input on success, -1 on failure. This takes its
|
|
own reference to `input`. */
|
|
static Py_ssize_t
|
|
_Unpickler_SetStringInput(UnpicklerObject *self, PyObject *input)
|
|
{
|
|
if (self->buffer.buf != NULL)
|
|
PyBuffer_Release(&self->buffer);
|
|
if (PyObject_GetBuffer(input, &self->buffer, PyBUF_CONTIG_RO) < 0)
|
|
return -1;
|
|
self->input_buffer = self->buffer.buf;
|
|
self->input_len = self->buffer.len;
|
|
self->next_read_idx = 0;
|
|
self->prefetched_idx = self->input_len;
|
|
return self->input_len;
|
|
}
|
|
|
|
static int
|
|
bad_readline(void)
|
|
{
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->UnpicklingError, "pickle data was truncated");
|
|
return -1;
|
|
}
|
|
|
|
/* Skip any consumed data that was only prefetched using peek() */
|
|
static int
|
|
_Unpickler_SkipConsumed(UnpicklerObject *self)
|
|
{
|
|
Py_ssize_t consumed;
|
|
PyObject *r;
|
|
|
|
consumed = self->next_read_idx - self->prefetched_idx;
|
|
if (consumed <= 0)
|
|
return 0;
|
|
|
|
assert(self->peek); /* otherwise we did something wrong */
|
|
/* This makes a useless copy... */
|
|
r = PyObject_CallFunction(self->read, "n", consumed);
|
|
if (r == NULL)
|
|
return -1;
|
|
Py_DECREF(r);
|
|
|
|
self->prefetched_idx = self->next_read_idx;
|
|
return 0;
|
|
}
|
|
|
|
static const Py_ssize_t READ_WHOLE_LINE = -1;
|
|
|
|
/* If reading from a file, we need to only pull the bytes we need, since there
|
|
may be multiple pickle objects arranged contiguously in the same input
|
|
buffer.
|
|
|
|
If `n` is READ_WHOLE_LINE, read a whole line. Otherwise, read up to `n`
|
|
bytes from the input stream/buffer.
|
|
|
|
Update the unpickler's input buffer with the newly-read data. Returns -1 on
|
|
failure; on success, returns the number of bytes read from the file.
|
|
|
|
On success, self->input_len will be 0; this is intentional so that when
|
|
unpickling from a file, the "we've run out of data" code paths will trigger,
|
|
causing the Unpickler to go back to the file for more data. Use the returned
|
|
size to tell you how much data you can process. */
|
|
static Py_ssize_t
|
|
_Unpickler_ReadFromFile(UnpicklerObject *self, Py_ssize_t n)
|
|
{
|
|
PyObject *data;
|
|
Py_ssize_t read_size;
|
|
|
|
assert(self->read != NULL);
|
|
|
|
if (_Unpickler_SkipConsumed(self) < 0)
|
|
return -1;
|
|
|
|
if (n == READ_WHOLE_LINE) {
|
|
data = PyObject_CallNoArgs(self->readline);
|
|
}
|
|
else {
|
|
PyObject *len;
|
|
/* Prefetch some data without advancing the file pointer, if possible */
|
|
if (self->peek && n < PREFETCH) {
|
|
len = PyLong_FromSsize_t(PREFETCH);
|
|
if (len == NULL)
|
|
return -1;
|
|
data = _Pickle_FastCall(self->peek, len);
|
|
if (data == NULL) {
|
|
if (!PyErr_ExceptionMatches(PyExc_NotImplementedError))
|
|
return -1;
|
|
/* peek() is probably not supported by the given file object */
|
|
PyErr_Clear();
|
|
Py_CLEAR(self->peek);
|
|
}
|
|
else {
|
|
read_size = _Unpickler_SetStringInput(self, data);
|
|
Py_DECREF(data);
|
|
self->prefetched_idx = 0;
|
|
if (n <= read_size)
|
|
return n;
|
|
}
|
|
}
|
|
len = PyLong_FromSsize_t(n);
|
|
if (len == NULL)
|
|
return -1;
|
|
data = _Pickle_FastCall(self->read, len);
|
|
}
|
|
if (data == NULL)
|
|
return -1;
|
|
|
|
read_size = _Unpickler_SetStringInput(self, data);
|
|
Py_DECREF(data);
|
|
return read_size;
|
|
}
|
|
|
|
/* Don't call it directly: use _Unpickler_Read() */
|
|
static Py_ssize_t
|
|
_Unpickler_ReadImpl(UnpicklerObject *self, char **s, Py_ssize_t n)
|
|
{
|
|
Py_ssize_t num_read;
|
|
|
|
*s = NULL;
|
|
if (self->next_read_idx > PY_SSIZE_T_MAX - n) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->UnpicklingError,
|
|
"read would overflow (invalid bytecode)");
|
|
return -1;
|
|
}
|
|
|
|
/* This case is handled by the _Unpickler_Read() macro for efficiency */
|
|
assert(self->next_read_idx + n > self->input_len);
|
|
|
|
if (!self->read)
|
|
return bad_readline();
|
|
|
|
/* Extend the buffer to satisfy desired size */
|
|
num_read = _Unpickler_ReadFromFile(self, n);
|
|
if (num_read < 0)
|
|
return -1;
|
|
if (num_read < n)
|
|
return bad_readline();
|
|
*s = self->input_buffer;
|
|
self->next_read_idx = n;
|
|
return n;
|
|
}
|
|
|
|
/* Read `n` bytes from the unpickler's data source, storing the result in `buf`.
|
|
*
|
|
* This should only be used for non-small data reads where potentially
|
|
* avoiding a copy is beneficial. This method does not try to prefetch
|
|
* more data into the input buffer.
|
|
*
|
|
* _Unpickler_Read() is recommended in most cases.
|
|
*/
|
|
static Py_ssize_t
|
|
_Unpickler_ReadInto(UnpicklerObject *self, char *buf, Py_ssize_t n)
|
|
{
|
|
assert(n != READ_WHOLE_LINE);
|
|
|
|
/* Read from available buffer data, if any */
|
|
Py_ssize_t in_buffer = self->input_len - self->next_read_idx;
|
|
if (in_buffer > 0) {
|
|
Py_ssize_t to_read = Py_MIN(in_buffer, n);
|
|
memcpy(buf, self->input_buffer + self->next_read_idx, to_read);
|
|
self->next_read_idx += to_read;
|
|
buf += to_read;
|
|
n -= to_read;
|
|
if (n == 0) {
|
|
/* Entire read was satisfied from buffer */
|
|
return n;
|
|
}
|
|
}
|
|
|
|
/* Read from file */
|
|
if (!self->read) {
|
|
/* We're unpickling memory, this means the input is truncated */
|
|
return bad_readline();
|
|
}
|
|
if (_Unpickler_SkipConsumed(self) < 0) {
|
|
return -1;
|
|
}
|
|
|
|
if (!self->readinto) {
|
|
/* readinto() not supported on file-like object, fall back to read()
|
|
* and copy into destination buffer (bpo-39681) */
|
|
PyObject* len = PyLong_FromSsize_t(n);
|
|
if (len == NULL) {
|
|
return -1;
|
|
}
|
|
PyObject* data = _Pickle_FastCall(self->read, len);
|
|
if (data == NULL) {
|
|
return -1;
|
|
}
|
|
if (!PyBytes_Check(data)) {
|
|
PyErr_Format(PyExc_ValueError,
|
|
"read() returned non-bytes object (%R)",
|
|
Py_TYPE(data));
|
|
Py_DECREF(data);
|
|
return -1;
|
|
}
|
|
Py_ssize_t read_size = PyBytes_GET_SIZE(data);
|
|
if (read_size < n) {
|
|
Py_DECREF(data);
|
|
return bad_readline();
|
|
}
|
|
memcpy(buf, PyBytes_AS_STRING(data), n);
|
|
Py_DECREF(data);
|
|
return n;
|
|
}
|
|
|
|
/* Call readinto() into user buffer */
|
|
PyObject *buf_obj = PyMemoryView_FromMemory(buf, n, PyBUF_WRITE);
|
|
if (buf_obj == NULL) {
|
|
return -1;
|
|
}
|
|
PyObject *read_size_obj = _Pickle_FastCall(self->readinto, buf_obj);
|
|
if (read_size_obj == NULL) {
|
|
return -1;
|
|
}
|
|
Py_ssize_t read_size = PyLong_AsSsize_t(read_size_obj);
|
|
Py_DECREF(read_size_obj);
|
|
|
|
if (read_size < 0) {
|
|
if (!PyErr_Occurred()) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"readinto() returned negative size");
|
|
}
|
|
return -1;
|
|
}
|
|
if (read_size < n) {
|
|
return bad_readline();
|
|
}
|
|
return n;
|
|
}
|
|
|
|
/* Read `n` bytes from the unpickler's data source, storing the result in `*s`.
|
|
|
|
This should be used for all data reads, rather than accessing the unpickler's
|
|
input buffer directly. This method deals correctly with reading from input
|
|
streams, which the input buffer doesn't deal with.
|
|
|
|
Note that when reading from a file-like object, self->next_read_idx won't
|
|
be updated (it should remain at 0 for the entire unpickling process). You
|
|
should use this function's return value to know how many bytes you can
|
|
consume.
|
|
|
|
Returns -1 (with an exception set) on failure. On success, return the
|
|
number of chars read. */
|
|
#define _Unpickler_Read(self, s, n) \
|
|
(((n) <= (self)->input_len - (self)->next_read_idx) \
|
|
? (*(s) = (self)->input_buffer + (self)->next_read_idx, \
|
|
(self)->next_read_idx += (n), \
|
|
(n)) \
|
|
: _Unpickler_ReadImpl(self, (s), (n)))
|
|
|
|
static Py_ssize_t
|
|
_Unpickler_CopyLine(UnpicklerObject *self, char *line, Py_ssize_t len,
|
|
char **result)
|
|
{
|
|
char *input_line = PyMem_Realloc(self->input_line, len + 1);
|
|
if (input_line == NULL) {
|
|
PyErr_NoMemory();
|
|
return -1;
|
|
}
|
|
|
|
memcpy(input_line, line, len);
|
|
input_line[len] = '\0';
|
|
self->input_line = input_line;
|
|
*result = self->input_line;
|
|
return len;
|
|
}
|
|
|
|
/* Read a line from the input stream/buffer. If we run off the end of the input
|
|
before hitting \n, raise an error.
|
|
|
|
Returns the number of chars read, or -1 on failure. */
|
|
static Py_ssize_t
|
|
_Unpickler_Readline(UnpicklerObject *self, char **result)
|
|
{
|
|
Py_ssize_t i, num_read;
|
|
|
|
for (i = self->next_read_idx; i < self->input_len; i++) {
|
|
if (self->input_buffer[i] == '\n') {
|
|
char *line_start = self->input_buffer + self->next_read_idx;
|
|
num_read = i - self->next_read_idx + 1;
|
|
self->next_read_idx = i + 1;
|
|
return _Unpickler_CopyLine(self, line_start, num_read, result);
|
|
}
|
|
}
|
|
if (!self->read)
|
|
return bad_readline();
|
|
|
|
num_read = _Unpickler_ReadFromFile(self, READ_WHOLE_LINE);
|
|
if (num_read < 0)
|
|
return -1;
|
|
if (num_read == 0 || self->input_buffer[num_read - 1] != '\n')
|
|
return bad_readline();
|
|
self->next_read_idx = num_read;
|
|
return _Unpickler_CopyLine(self, self->input_buffer, num_read, result);
|
|
}
|
|
|
|
/* Returns -1 (with an exception set) on failure, 0 on success. The memo array
|
|
will be modified in place. */
|
|
static int
|
|
_Unpickler_ResizeMemoList(UnpicklerObject *self, size_t new_size)
|
|
{
|
|
size_t i;
|
|
|
|
assert(new_size > self->memo_size);
|
|
|
|
PyObject **memo_new = self->memo;
|
|
PyMem_RESIZE(memo_new, PyObject *, new_size);
|
|
if (memo_new == NULL) {
|
|
PyErr_NoMemory();
|
|
return -1;
|
|
}
|
|
self->memo = memo_new;
|
|
for (i = self->memo_size; i < new_size; i++)
|
|
self->memo[i] = NULL;
|
|
self->memo_size = new_size;
|
|
return 0;
|
|
}
|
|
|
|
/* Returns NULL if idx is out of bounds. */
|
|
static PyObject *
|
|
_Unpickler_MemoGet(UnpicklerObject *self, size_t idx)
|
|
{
|
|
if (idx >= self->memo_size)
|
|
return NULL;
|
|
|
|
return self->memo[idx];
|
|
}
|
|
|
|
/* Returns -1 (with an exception set) on failure, 0 on success.
|
|
This takes its own reference to `value`. */
|
|
static int
|
|
_Unpickler_MemoPut(UnpicklerObject *self, size_t idx, PyObject *value)
|
|
{
|
|
PyObject *old_item;
|
|
|
|
if (idx >= self->memo_size) {
|
|
if (_Unpickler_ResizeMemoList(self, idx * 2) < 0)
|
|
return -1;
|
|
assert(idx < self->memo_size);
|
|
}
|
|
Py_INCREF(value);
|
|
old_item = self->memo[idx];
|
|
self->memo[idx] = value;
|
|
if (old_item != NULL) {
|
|
Py_DECREF(old_item);
|
|
}
|
|
else {
|
|
self->memo_len++;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static PyObject **
|
|
_Unpickler_NewMemo(Py_ssize_t new_size)
|
|
{
|
|
PyObject **memo = PyMem_NEW(PyObject *, new_size);
|
|
if (memo == NULL) {
|
|
PyErr_NoMemory();
|
|
return NULL;
|
|
}
|
|
memset(memo, 0, new_size * sizeof(PyObject *));
|
|
return memo;
|
|
}
|
|
|
|
/* Free the unpickler's memo, taking care to decref any items left in it. */
|
|
static void
|
|
_Unpickler_MemoCleanup(UnpicklerObject *self)
|
|
{
|
|
Py_ssize_t i;
|
|
PyObject **memo = self->memo;
|
|
|
|
if (self->memo == NULL)
|
|
return;
|
|
self->memo = NULL;
|
|
i = self->memo_size;
|
|
while (--i >= 0) {
|
|
Py_XDECREF(memo[i]);
|
|
}
|
|
PyMem_FREE(memo);
|
|
}
|
|
|
|
static UnpicklerObject *
|
|
_Unpickler_New(void)
|
|
{
|
|
UnpicklerObject *self;
|
|
|
|
self = PyObject_GC_New(UnpicklerObject, &Unpickler_Type);
|
|
if (self == NULL)
|
|
return NULL;
|
|
|
|
self->pers_func = NULL;
|
|
self->input_buffer = NULL;
|
|
self->input_line = NULL;
|
|
self->input_len = 0;
|
|
self->next_read_idx = 0;
|
|
self->prefetched_idx = 0;
|
|
self->read = NULL;
|
|
self->readinto = NULL;
|
|
self->readline = NULL;
|
|
self->peek = NULL;
|
|
self->buffers = NULL;
|
|
self->encoding = NULL;
|
|
self->errors = NULL;
|
|
self->marks = NULL;
|
|
self->num_marks = 0;
|
|
self->marks_size = 0;
|
|
self->proto = 0;
|
|
self->fix_imports = 0;
|
|
memset(&self->buffer, 0, sizeof(Py_buffer));
|
|
self->memo_size = 32;
|
|
self->memo_len = 0;
|
|
self->memo = _Unpickler_NewMemo(self->memo_size);
|
|
self->stack = (Pdata *)Pdata_New();
|
|
|
|
if (self->memo == NULL || self->stack == NULL) {
|
|
Py_DECREF(self);
|
|
return NULL;
|
|
}
|
|
|
|
PyObject_GC_Track(self);
|
|
return self;
|
|
}
|
|
|
|
/* Returns -1 (with an exception set) on failure, 0 on success. This may
|
|
be called once on a freshly created Unpickler. */
|
|
static int
|
|
_Unpickler_SetInputStream(UnpicklerObject *self, PyObject *file)
|
|
{
|
|
_Py_IDENTIFIER(peek);
|
|
_Py_IDENTIFIER(read);
|
|
_Py_IDENTIFIER(readinto);
|
|
_Py_IDENTIFIER(readline);
|
|
|
|
/* Optional file methods */
|
|
if (_PyObject_LookupAttrId(file, &PyId_peek, &self->peek) < 0) {
|
|
return -1;
|
|
}
|
|
if (_PyObject_LookupAttrId(file, &PyId_readinto, &self->readinto) < 0) {
|
|
return -1;
|
|
}
|
|
(void)_PyObject_LookupAttrId(file, &PyId_read, &self->read);
|
|
(void)_PyObject_LookupAttrId(file, &PyId_readline, &self->readline);
|
|
if (!self->readline || !self->read) {
|
|
if (!PyErr_Occurred()) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"file must have 'read' and 'readline' attributes");
|
|
}
|
|
Py_CLEAR(self->read);
|
|
Py_CLEAR(self->readinto);
|
|
Py_CLEAR(self->readline);
|
|
Py_CLEAR(self->peek);
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/* Returns -1 (with an exception set) on failure, 0 on success. This may
|
|
be called once on a freshly created Unpickler. */
|
|
static int
|
|
_Unpickler_SetInputEncoding(UnpicklerObject *self,
|
|
const char *encoding,
|
|
const char *errors)
|
|
{
|
|
if (encoding == NULL)
|
|
encoding = "ASCII";
|
|
if (errors == NULL)
|
|
errors = "strict";
|
|
|
|
self->encoding = _PyMem_Strdup(encoding);
|
|
self->errors = _PyMem_Strdup(errors);
|
|
if (self->encoding == NULL || self->errors == NULL) {
|
|
PyErr_NoMemory();
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/* Returns -1 (with an exception set) on failure, 0 on success. This may
|
|
be called once on a freshly created Unpickler. */
|
|
static int
|
|
_Unpickler_SetBuffers(UnpicklerObject *self, PyObject *buffers)
|
|
{
|
|
if (buffers == NULL || buffers == Py_None) {
|
|
self->buffers = NULL;
|
|
}
|
|
else {
|
|
self->buffers = PyObject_GetIter(buffers);
|
|
if (self->buffers == NULL) {
|
|
return -1;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/* Generate a GET opcode for an object stored in the memo. */
|
|
static int
|
|
memo_get(PicklerObject *self, PyObject *key)
|
|
{
|
|
Py_ssize_t *value;
|
|
char pdata[30];
|
|
Py_ssize_t len;
|
|
|
|
value = PyMemoTable_Get(self->memo, key);
|
|
if (value == NULL) {
|
|
PyErr_SetObject(PyExc_KeyError, key);
|
|
return -1;
|
|
}
|
|
|
|
if (!self->bin) {
|
|
pdata[0] = GET;
|
|
PyOS_snprintf(pdata + 1, sizeof(pdata) - 1,
|
|
"%zd\n", *value);
|
|
len = strlen(pdata);
|
|
}
|
|
else {
|
|
if (*value < 256) {
|
|
pdata[0] = BINGET;
|
|
pdata[1] = (unsigned char)(*value & 0xff);
|
|
len = 2;
|
|
}
|
|
else if ((size_t)*value <= 0xffffffffUL) {
|
|
pdata[0] = LONG_BINGET;
|
|
pdata[1] = (unsigned char)(*value & 0xff);
|
|
pdata[2] = (unsigned char)((*value >> 8) & 0xff);
|
|
pdata[3] = (unsigned char)((*value >> 16) & 0xff);
|
|
pdata[4] = (unsigned char)((*value >> 24) & 0xff);
|
|
len = 5;
|
|
}
|
|
else { /* unlikely */
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->PicklingError,
|
|
"memo id too large for LONG_BINGET");
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
if (_Pickler_Write(self, pdata, len) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Store an object in the memo, assign it a new unique ID based on the number
|
|
of objects currently stored in the memo and generate a PUT opcode. */
|
|
static int
|
|
memo_put(PicklerObject *self, PyObject *obj)
|
|
{
|
|
char pdata[30];
|
|
Py_ssize_t len;
|
|
Py_ssize_t idx;
|
|
|
|
const char memoize_op = MEMOIZE;
|
|
|
|
if (self->fast)
|
|
return 0;
|
|
|
|
idx = PyMemoTable_Size(self->memo);
|
|
if (PyMemoTable_Set(self->memo, obj, idx) < 0)
|
|
return -1;
|
|
|
|
if (self->proto >= 4) {
|
|
if (_Pickler_Write(self, &memoize_op, 1) < 0)
|
|
return -1;
|
|
return 0;
|
|
}
|
|
else if (!self->bin) {
|
|
pdata[0] = PUT;
|
|
PyOS_snprintf(pdata + 1, sizeof(pdata) - 1,
|
|
"%zd\n", idx);
|
|
len = strlen(pdata);
|
|
}
|
|
else {
|
|
if (idx < 256) {
|
|
pdata[0] = BINPUT;
|
|
pdata[1] = (unsigned char)idx;
|
|
len = 2;
|
|
}
|
|
else if ((size_t)idx <= 0xffffffffUL) {
|
|
pdata[0] = LONG_BINPUT;
|
|
pdata[1] = (unsigned char)(idx & 0xff);
|
|
pdata[2] = (unsigned char)((idx >> 8) & 0xff);
|
|
pdata[3] = (unsigned char)((idx >> 16) & 0xff);
|
|
pdata[4] = (unsigned char)((idx >> 24) & 0xff);
|
|
len = 5;
|
|
}
|
|
else { /* unlikely */
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->PicklingError,
|
|
"memo id too large for LONG_BINPUT");
|
|
return -1;
|
|
}
|
|
}
|
|
if (_Pickler_Write(self, pdata, len) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static PyObject *
|
|
get_dotted_path(PyObject *obj, PyObject *name)
|
|
{
|
|
_Py_static_string(PyId_dot, ".");
|
|
PyObject *dotted_path;
|
|
Py_ssize_t i, n;
|
|
|
|
dotted_path = PyUnicode_Split(name, _PyUnicode_FromId(&PyId_dot), -1);
|
|
if (dotted_path == NULL)
|
|
return NULL;
|
|
n = PyList_GET_SIZE(dotted_path);
|
|
assert(n >= 1);
|
|
for (i = 0; i < n; i++) {
|
|
PyObject *subpath = PyList_GET_ITEM(dotted_path, i);
|
|
if (_PyUnicode_EqualToASCIIString(subpath, "<locals>")) {
|
|
if (obj == NULL)
|
|
PyErr_Format(PyExc_AttributeError,
|
|
"Can't pickle local object %R", name);
|
|
else
|
|
PyErr_Format(PyExc_AttributeError,
|
|
"Can't pickle local attribute %R on %R", name, obj);
|
|
Py_DECREF(dotted_path);
|
|
return NULL;
|
|
}
|
|
}
|
|
return dotted_path;
|
|
}
|
|
|
|
static PyObject *
|
|
get_deep_attribute(PyObject *obj, PyObject *names, PyObject **pparent)
|
|
{
|
|
Py_ssize_t i, n;
|
|
PyObject *parent = NULL;
|
|
|
|
assert(PyList_CheckExact(names));
|
|
Py_INCREF(obj);
|
|
n = PyList_GET_SIZE(names);
|
|
for (i = 0; i < n; i++) {
|
|
PyObject *name = PyList_GET_ITEM(names, i);
|
|
Py_XDECREF(parent);
|
|
parent = obj;
|
|
(void)_PyObject_LookupAttr(parent, name, &obj);
|
|
if (obj == NULL) {
|
|
Py_DECREF(parent);
|
|
return NULL;
|
|
}
|
|
}
|
|
if (pparent != NULL)
|
|
*pparent = parent;
|
|
else
|
|
Py_XDECREF(parent);
|
|
return obj;
|
|
}
|
|
|
|
|
|
static PyObject *
|
|
getattribute(PyObject *obj, PyObject *name, int allow_qualname)
|
|
{
|
|
PyObject *dotted_path, *attr;
|
|
|
|
if (allow_qualname) {
|
|
dotted_path = get_dotted_path(obj, name);
|
|
if (dotted_path == NULL)
|
|
return NULL;
|
|
attr = get_deep_attribute(obj, dotted_path, NULL);
|
|
Py_DECREF(dotted_path);
|
|
}
|
|
else {
|
|
(void)_PyObject_LookupAttr(obj, name, &attr);
|
|
}
|
|
if (attr == NULL && !PyErr_Occurred()) {
|
|
PyErr_Format(PyExc_AttributeError,
|
|
"Can't get attribute %R on %R", name, obj);
|
|
}
|
|
return attr;
|
|
}
|
|
|
|
static int
|
|
_checkmodule(PyObject *module_name, PyObject *module,
|
|
PyObject *global, PyObject *dotted_path)
|
|
{
|
|
if (module == Py_None) {
|
|
return -1;
|
|
}
|
|
if (PyUnicode_Check(module_name) &&
|
|
_PyUnicode_EqualToASCIIString(module_name, "__main__")) {
|
|
return -1;
|
|
}
|
|
|
|
PyObject *candidate = get_deep_attribute(module, dotted_path, NULL);
|
|
if (candidate == NULL) {
|
|
return -1;
|
|
}
|
|
if (candidate != global) {
|
|
Py_DECREF(candidate);
|
|
return -1;
|
|
}
|
|
Py_DECREF(candidate);
|
|
return 0;
|
|
}
|
|
|
|
static PyObject *
|
|
whichmodule(PyObject *global, PyObject *dotted_path)
|
|
{
|
|
PyObject *module_name;
|
|
PyObject *module = NULL;
|
|
Py_ssize_t i;
|
|
PyObject *modules;
|
|
_Py_IDENTIFIER(__module__);
|
|
_Py_IDENTIFIER(modules);
|
|
_Py_IDENTIFIER(__main__);
|
|
|
|
if (_PyObject_LookupAttrId(global, &PyId___module__, &module_name) < 0) {
|
|
return NULL;
|
|
}
|
|
if (module_name) {
|
|
/* In some rare cases (e.g., bound methods of extension types),
|
|
__module__ can be None. If it is so, then search sys.modules for
|
|
the module of global. */
|
|
if (module_name != Py_None)
|
|
return module_name;
|
|
Py_CLEAR(module_name);
|
|
}
|
|
assert(module_name == NULL);
|
|
|
|
/* Fallback on walking sys.modules */
|
|
modules = _PySys_GetObjectId(&PyId_modules);
|
|
if (modules == NULL) {
|
|
PyErr_SetString(PyExc_RuntimeError, "unable to get sys.modules");
|
|
return NULL;
|
|
}
|
|
if (PyDict_CheckExact(modules)) {
|
|
i = 0;
|
|
while (PyDict_Next(modules, &i, &module_name, &module)) {
|
|
if (_checkmodule(module_name, module, global, dotted_path) == 0) {
|
|
Py_INCREF(module_name);
|
|
return module_name;
|
|
}
|
|
if (PyErr_Occurred()) {
|
|
return NULL;
|
|
}
|
|
}
|
|
}
|
|
else {
|
|
PyObject *iterator = PyObject_GetIter(modules);
|
|
if (iterator == NULL) {
|
|
return NULL;
|
|
}
|
|
while ((module_name = PyIter_Next(iterator))) {
|
|
module = PyObject_GetItem(modules, module_name);
|
|
if (module == NULL) {
|
|
Py_DECREF(module_name);
|
|
Py_DECREF(iterator);
|
|
return NULL;
|
|
}
|
|
if (_checkmodule(module_name, module, global, dotted_path) == 0) {
|
|
Py_DECREF(module);
|
|
Py_DECREF(iterator);
|
|
return module_name;
|
|
}
|
|
Py_DECREF(module);
|
|
Py_DECREF(module_name);
|
|
if (PyErr_Occurred()) {
|
|
Py_DECREF(iterator);
|
|
return NULL;
|
|
}
|
|
}
|
|
Py_DECREF(iterator);
|
|
}
|
|
|
|
/* If no module is found, use __main__. */
|
|
module_name = _PyUnicode_FromId(&PyId___main__);
|
|
Py_XINCREF(module_name);
|
|
return module_name;
|
|
}
|
|
|
|
/* fast_save_enter() and fast_save_leave() are guards against recursive
|
|
objects when Pickler is used with the "fast mode" (i.e., with object
|
|
memoization disabled). If the nesting of a list or dict object exceed
|
|
FAST_NESTING_LIMIT, these guards will start keeping an internal
|
|
reference to the seen list or dict objects and check whether these objects
|
|
are recursive. These are not strictly necessary, since save() has a
|
|
hard-coded recursion limit, but they give a nicer error message than the
|
|
typical RuntimeError. */
|
|
static int
|
|
fast_save_enter(PicklerObject *self, PyObject *obj)
|
|
{
|
|
/* if fast_nesting < 0, we're doing an error exit. */
|
|
if (++self->fast_nesting >= FAST_NESTING_LIMIT) {
|
|
PyObject *key = NULL;
|
|
if (self->fast_memo == NULL) {
|
|
self->fast_memo = PyDict_New();
|
|
if (self->fast_memo == NULL) {
|
|
self->fast_nesting = -1;
|
|
return 0;
|
|
}
|
|
}
|
|
key = PyLong_FromVoidPtr(obj);
|
|
if (key == NULL) {
|
|
self->fast_nesting = -1;
|
|
return 0;
|
|
}
|
|
int r = PyDict_Contains(self->fast_memo, key);
|
|
if (r > 0) {
|
|
PyErr_Format(PyExc_ValueError,
|
|
"fast mode: can't pickle cyclic objects "
|
|
"including object type %.200s at %p",
|
|
Py_TYPE(obj)->tp_name, obj);
|
|
}
|
|
else if (r == 0) {
|
|
r = PyDict_SetItem(self->fast_memo, key, Py_None);
|
|
}
|
|
Py_DECREF(key);
|
|
if (r != 0) {
|
|
self->fast_nesting = -1;
|
|
return 0;
|
|
}
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
static int
|
|
fast_save_leave(PicklerObject *self, PyObject *obj)
|
|
{
|
|
if (self->fast_nesting-- >= FAST_NESTING_LIMIT) {
|
|
PyObject *key = PyLong_FromVoidPtr(obj);
|
|
if (key == NULL)
|
|
return 0;
|
|
if (PyDict_DelItem(self->fast_memo, key) < 0) {
|
|
Py_DECREF(key);
|
|
return 0;
|
|
}
|
|
Py_DECREF(key);
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
static int
|
|
save_none(PicklerObject *self, PyObject *obj)
|
|
{
|
|
const char none_op = NONE;
|
|
if (_Pickler_Write(self, &none_op, 1) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
save_bool(PicklerObject *self, PyObject *obj)
|
|
{
|
|
if (self->proto >= 2) {
|
|
const char bool_op = (obj == Py_True) ? NEWTRUE : NEWFALSE;
|
|
if (_Pickler_Write(self, &bool_op, 1) < 0)
|
|
return -1;
|
|
}
|
|
else {
|
|
/* These aren't opcodes -- they're ways to pickle bools before protocol 2
|
|
* so that unpicklers written before bools were introduced unpickle them
|
|
* as ints, but unpicklers after can recognize that bools were intended.
|
|
* Note that protocol 2 added direct ways to pickle bools.
|
|
*/
|
|
const char *bool_str = (obj == Py_True) ? "I01\n" : "I00\n";
|
|
if (_Pickler_Write(self, bool_str, strlen(bool_str)) < 0)
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
save_long(PicklerObject *self, PyObject *obj)
|
|
{
|
|
PyObject *repr = NULL;
|
|
Py_ssize_t size;
|
|
long val;
|
|
int overflow;
|
|
int status = 0;
|
|
|
|
val= PyLong_AsLongAndOverflow(obj, &overflow);
|
|
if (!overflow && (sizeof(long) <= 4 ||
|
|
(val <= 0x7fffffffL && val >= (-0x7fffffffL - 1))))
|
|
{
|
|
/* result fits in a signed 4-byte integer.
|
|
|
|
Note: we can't use -0x80000000L in the above condition because some
|
|
compilers (e.g., MSVC) will promote 0x80000000L to an unsigned type
|
|
before applying the unary minus when sizeof(long) <= 4. The
|
|
resulting value stays unsigned which is commonly not what we want,
|
|
so MSVC happily warns us about it. However, that result would have
|
|
been fine because we guard for sizeof(long) <= 4 which turns the
|
|
condition true in that particular case. */
|
|
char pdata[32];
|
|
Py_ssize_t len = 0;
|
|
|
|
if (self->bin) {
|
|
pdata[1] = (unsigned char)(val & 0xff);
|
|
pdata[2] = (unsigned char)((val >> 8) & 0xff);
|
|
pdata[3] = (unsigned char)((val >> 16) & 0xff);
|
|
pdata[4] = (unsigned char)((val >> 24) & 0xff);
|
|
|
|
if ((pdata[4] != 0) || (pdata[3] != 0)) {
|
|
pdata[0] = BININT;
|
|
len = 5;
|
|
}
|
|
else if (pdata[2] != 0) {
|
|
pdata[0] = BININT2;
|
|
len = 3;
|
|
}
|
|
else {
|
|
pdata[0] = BININT1;
|
|
len = 2;
|
|
}
|
|
}
|
|
else {
|
|
sprintf(pdata, "%c%ld\n", INT, val);
|
|
len = strlen(pdata);
|
|
}
|
|
if (_Pickler_Write(self, pdata, len) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
assert(!PyErr_Occurred());
|
|
|
|
if (self->proto >= 2) {
|
|
/* Linear-time pickling. */
|
|
size_t nbits;
|
|
size_t nbytes;
|
|
unsigned char *pdata;
|
|
char header[5];
|
|
int i;
|
|
int sign = _PyLong_Sign(obj);
|
|
|
|
if (sign == 0) {
|
|
header[0] = LONG1;
|
|
header[1] = 0; /* It's 0 -- an empty bytestring. */
|
|
if (_Pickler_Write(self, header, 2) < 0)
|
|
goto error;
|
|
return 0;
|
|
}
|
|
nbits = _PyLong_NumBits(obj);
|
|
if (nbits == (size_t)-1 && PyErr_Occurred())
|
|
goto error;
|
|
/* How many bytes do we need? There are nbits >> 3 full
|
|
* bytes of data, and nbits & 7 leftover bits. If there
|
|
* are any leftover bits, then we clearly need another
|
|
* byte. What's not so obvious is that we *probably*
|
|
* need another byte even if there aren't any leftovers:
|
|
* the most-significant bit of the most-significant byte
|
|
* acts like a sign bit, and it's usually got a sense
|
|
* opposite of the one we need. The exception is ints
|
|
* of the form -(2**(8*j-1)) for j > 0. Such an int is
|
|
* its own 256's-complement, so has the right sign bit
|
|
* even without the extra byte. That's a pain to check
|
|
* for in advance, though, so we always grab an extra
|
|
* byte at the start, and cut it back later if possible.
|
|
*/
|
|
nbytes = (nbits >> 3) + 1;
|
|
if (nbytes > 0x7fffffffL) {
|
|
PyErr_SetString(PyExc_OverflowError,
|
|
"int too large to pickle");
|
|
goto error;
|
|
}
|
|
repr = PyBytes_FromStringAndSize(NULL, (Py_ssize_t)nbytes);
|
|
if (repr == NULL)
|
|
goto error;
|
|
pdata = (unsigned char *)PyBytes_AS_STRING(repr);
|
|
i = _PyLong_AsByteArray((PyLongObject *)obj,
|
|
pdata, nbytes,
|
|
1 /* little endian */ , 1 /* signed */ );
|
|
if (i < 0)
|
|
goto error;
|
|
/* If the int is negative, this may be a byte more than
|
|
* needed. This is so iff the MSB is all redundant sign
|
|
* bits.
|
|
*/
|
|
if (sign < 0 &&
|
|
nbytes > 1 &&
|
|
pdata[nbytes - 1] == 0xff &&
|
|
(pdata[nbytes - 2] & 0x80) != 0) {
|
|
nbytes--;
|
|
}
|
|
|
|
if (nbytes < 256) {
|
|
header[0] = LONG1;
|
|
header[1] = (unsigned char)nbytes;
|
|
size = 2;
|
|
}
|
|
else {
|
|
header[0] = LONG4;
|
|
size = (Py_ssize_t) nbytes;
|
|
for (i = 1; i < 5; i++) {
|
|
header[i] = (unsigned char)(size & 0xff);
|
|
size >>= 8;
|
|
}
|
|
size = 5;
|
|
}
|
|
if (_Pickler_Write(self, header, size) < 0 ||
|
|
_Pickler_Write(self, (char *)pdata, (int)nbytes) < 0)
|
|
goto error;
|
|
}
|
|
else {
|
|
const char long_op = LONG;
|
|
const char *string;
|
|
|
|
/* proto < 2: write the repr and newline. This is quadratic-time (in
|
|
the number of digits), in both directions. We add a trailing 'L'
|
|
to the repr, for compatibility with Python 2.x. */
|
|
|
|
repr = PyObject_Repr(obj);
|
|
if (repr == NULL)
|
|
goto error;
|
|
|
|
string = PyUnicode_AsUTF8AndSize(repr, &size);
|
|
if (string == NULL)
|
|
goto error;
|
|
|
|
if (_Pickler_Write(self, &long_op, 1) < 0 ||
|
|
_Pickler_Write(self, string, size) < 0 ||
|
|
_Pickler_Write(self, "L\n", 2) < 0)
|
|
goto error;
|
|
}
|
|
|
|
if (0) {
|
|
error:
|
|
status = -1;
|
|
}
|
|
Py_XDECREF(repr);
|
|
|
|
return status;
|
|
}
|
|
|
|
static int
|
|
save_float(PicklerObject *self, PyObject *obj)
|
|
{
|
|
double x = PyFloat_AS_DOUBLE((PyFloatObject *)obj);
|
|
|
|
if (self->bin) {
|
|
char pdata[9];
|
|
pdata[0] = BINFLOAT;
|
|
if (_PyFloat_Pack8(x, (unsigned char *)&pdata[1], 0) < 0)
|
|
return -1;
|
|
if (_Pickler_Write(self, pdata, 9) < 0)
|
|
return -1;
|
|
}
|
|
else {
|
|
int result = -1;
|
|
char *buf = NULL;
|
|
char op = FLOAT;
|
|
|
|
if (_Pickler_Write(self, &op, 1) < 0)
|
|
goto done;
|
|
|
|
buf = PyOS_double_to_string(x, 'r', 0, Py_DTSF_ADD_DOT_0, NULL);
|
|
if (!buf) {
|
|
PyErr_NoMemory();
|
|
goto done;
|
|
}
|
|
|
|
if (_Pickler_Write(self, buf, strlen(buf)) < 0)
|
|
goto done;
|
|
|
|
if (_Pickler_Write(self, "\n", 1) < 0)
|
|
goto done;
|
|
|
|
result = 0;
|
|
done:
|
|
PyMem_Free(buf);
|
|
return result;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Perform direct write of the header and payload of the binary object.
|
|
|
|
The large contiguous data is written directly into the underlying file
|
|
object, bypassing the output_buffer of the Pickler. We intentionally
|
|
do not insert a protocol 4 frame opcode to make it possible to optimize
|
|
file.read calls in the loader.
|
|
*/
|
|
static int
|
|
_Pickler_write_bytes(PicklerObject *self,
|
|
const char *header, Py_ssize_t header_size,
|
|
const char *data, Py_ssize_t data_size,
|
|
PyObject *payload)
|
|
{
|
|
int bypass_buffer = (data_size >= FRAME_SIZE_TARGET);
|
|
int framing = self->framing;
|
|
|
|
if (bypass_buffer) {
|
|
assert(self->output_buffer != NULL);
|
|
/* Commit the previous frame. */
|
|
if (_Pickler_CommitFrame(self)) {
|
|
return -1;
|
|
}
|
|
/* Disable framing temporarily */
|
|
self->framing = 0;
|
|
}
|
|
|
|
if (_Pickler_Write(self, header, header_size) < 0) {
|
|
return -1;
|
|
}
|
|
|
|
if (bypass_buffer && self->write != NULL) {
|
|
/* Bypass the in-memory buffer to directly stream large data
|
|
into the underlying file object. */
|
|
PyObject *result, *mem = NULL;
|
|
/* Dump the output buffer to the file. */
|
|
if (_Pickler_FlushToFile(self) < 0) {
|
|
return -1;
|
|
}
|
|
|
|
/* Stream write the payload into the file without going through the
|
|
output buffer. */
|
|
if (payload == NULL) {
|
|
/* TODO: It would be better to use a memoryview with a linked
|
|
original string if this is possible. */
|
|
payload = mem = PyBytes_FromStringAndSize(data, data_size);
|
|
if (payload == NULL) {
|
|
return -1;
|
|
}
|
|
}
|
|
result = PyObject_CallOneArg(self->write, payload);
|
|
Py_XDECREF(mem);
|
|
if (result == NULL) {
|
|
return -1;
|
|
}
|
|
Py_DECREF(result);
|
|
|
|
/* Reinitialize the buffer for subsequent calls to _Pickler_Write. */
|
|
if (_Pickler_ClearBuffer(self) < 0) {
|
|
return -1;
|
|
}
|
|
}
|
|
else {
|
|
if (_Pickler_Write(self, data, data_size) < 0) {
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
/* Re-enable framing for subsequent calls to _Pickler_Write. */
|
|
self->framing = framing;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
_save_bytes_data(PicklerObject *self, PyObject *obj, const char *data,
|
|
Py_ssize_t size)
|
|
{
|
|
assert(self->proto >= 3);
|
|
|
|
char header[9];
|
|
Py_ssize_t len;
|
|
|
|
if (size < 0)
|
|
return -1;
|
|
|
|
if (size <= 0xff) {
|
|
header[0] = SHORT_BINBYTES;
|
|
header[1] = (unsigned char)size;
|
|
len = 2;
|
|
}
|
|
else if ((size_t)size <= 0xffffffffUL) {
|
|
header[0] = BINBYTES;
|
|
header[1] = (unsigned char)(size & 0xff);
|
|
header[2] = (unsigned char)((size >> 8) & 0xff);
|
|
header[3] = (unsigned char)((size >> 16) & 0xff);
|
|
header[4] = (unsigned char)((size >> 24) & 0xff);
|
|
len = 5;
|
|
}
|
|
else if (self->proto >= 4) {
|
|
header[0] = BINBYTES8;
|
|
_write_size64(header + 1, size);
|
|
len = 9;
|
|
}
|
|
else {
|
|
PyErr_SetString(PyExc_OverflowError,
|
|
"serializing a bytes object larger than 4 GiB "
|
|
"requires pickle protocol 4 or higher");
|
|
return -1;
|
|
}
|
|
|
|
if (_Pickler_write_bytes(self, header, len, data, size, obj) < 0) {
|
|
return -1;
|
|
}
|
|
|
|
if (memo_put(self, obj) < 0) {
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
save_bytes(PicklerObject *self, PyObject *obj)
|
|
{
|
|
if (self->proto < 3) {
|
|
/* Older pickle protocols do not have an opcode for pickling bytes
|
|
objects. Therefore, we need to fake the copy protocol (i.e.,
|
|
the __reduce__ method) to permit bytes object unpickling.
|
|
|
|
Here we use a hack to be compatible with Python 2. Since in Python
|
|
2 'bytes' is just an alias for 'str' (which has different
|
|
parameters than the actual bytes object), we use codecs.encode
|
|
to create the appropriate 'str' object when unpickled using
|
|
Python 2 *and* the appropriate 'bytes' object when unpickled
|
|
using Python 3. Again this is a hack and we don't need to do this
|
|
with newer protocols. */
|
|
PyObject *reduce_value;
|
|
int status;
|
|
|
|
if (PyBytes_GET_SIZE(obj) == 0) {
|
|
reduce_value = Py_BuildValue("(O())", (PyObject*)&PyBytes_Type);
|
|
}
|
|
else {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyObject *unicode_str =
|
|
PyUnicode_DecodeLatin1(PyBytes_AS_STRING(obj),
|
|
PyBytes_GET_SIZE(obj),
|
|
"strict");
|
|
_Py_IDENTIFIER(latin1);
|
|
|
|
if (unicode_str == NULL)
|
|
return -1;
|
|
reduce_value = Py_BuildValue("(O(OO))",
|
|
st->codecs_encode, unicode_str,
|
|
_PyUnicode_FromId(&PyId_latin1));
|
|
Py_DECREF(unicode_str);
|
|
}
|
|
|
|
if (reduce_value == NULL)
|
|
return -1;
|
|
|
|
/* save_reduce() will memoize the object automatically. */
|
|
status = save_reduce(self, reduce_value, obj);
|
|
Py_DECREF(reduce_value);
|
|
return status;
|
|
}
|
|
else {
|
|
return _save_bytes_data(self, obj, PyBytes_AS_STRING(obj),
|
|
PyBytes_GET_SIZE(obj));
|
|
}
|
|
}
|
|
|
|
static int
|
|
_save_bytearray_data(PicklerObject *self, PyObject *obj, const char *data,
|
|
Py_ssize_t size)
|
|
{
|
|
assert(self->proto >= 5);
|
|
|
|
char header[9];
|
|
Py_ssize_t len;
|
|
|
|
if (size < 0)
|
|
return -1;
|
|
|
|
header[0] = BYTEARRAY8;
|
|
_write_size64(header + 1, size);
|
|
len = 9;
|
|
|
|
if (_Pickler_write_bytes(self, header, len, data, size, obj) < 0) {
|
|
return -1;
|
|
}
|
|
|
|
if (memo_put(self, obj) < 0) {
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
save_bytearray(PicklerObject *self, PyObject *obj)
|
|
{
|
|
if (self->proto < 5) {
|
|
/* Older pickle protocols do not have an opcode for pickling
|
|
* bytearrays. */
|
|
PyObject *reduce_value = NULL;
|
|
int status;
|
|
|
|
if (PyByteArray_GET_SIZE(obj) == 0) {
|
|
reduce_value = Py_BuildValue("(O())",
|
|
(PyObject *) &PyByteArray_Type);
|
|
}
|
|
else {
|
|
PyObject *bytes_obj = PyBytes_FromObject(obj);
|
|
if (bytes_obj != NULL) {
|
|
reduce_value = Py_BuildValue("(O(O))",
|
|
(PyObject *) &PyByteArray_Type,
|
|
bytes_obj);
|
|
Py_DECREF(bytes_obj);
|
|
}
|
|
}
|
|
if (reduce_value == NULL)
|
|
return -1;
|
|
|
|
/* save_reduce() will memoize the object automatically. */
|
|
status = save_reduce(self, reduce_value, obj);
|
|
Py_DECREF(reduce_value);
|
|
return status;
|
|
}
|
|
else {
|
|
return _save_bytearray_data(self, obj, PyByteArray_AS_STRING(obj),
|
|
PyByteArray_GET_SIZE(obj));
|
|
}
|
|
}
|
|
|
|
static int
|
|
save_picklebuffer(PicklerObject *self, PyObject *obj)
|
|
{
|
|
if (self->proto < 5) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->PicklingError,
|
|
"PickleBuffer can only pickled with protocol >= 5");
|
|
return -1;
|
|
}
|
|
const Py_buffer* view = PyPickleBuffer_GetBuffer(obj);
|
|
if (view == NULL) {
|
|
return -1;
|
|
}
|
|
if (view->suboffsets != NULL || !PyBuffer_IsContiguous(view, 'A')) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->PicklingError,
|
|
"PickleBuffer can not be pickled when "
|
|
"pointing to a non-contiguous buffer");
|
|
return -1;
|
|
}
|
|
int in_band = 1;
|
|
if (self->buffer_callback != NULL) {
|
|
PyObject *ret = PyObject_CallOneArg(self->buffer_callback, obj);
|
|
if (ret == NULL) {
|
|
return -1;
|
|
}
|
|
in_band = PyObject_IsTrue(ret);
|
|
Py_DECREF(ret);
|
|
if (in_band == -1) {
|
|
return -1;
|
|
}
|
|
}
|
|
if (in_band) {
|
|
/* Write data in-band */
|
|
if (view->readonly) {
|
|
return _save_bytes_data(self, obj, (const char*) view->buf,
|
|
view->len);
|
|
}
|
|
else {
|
|
return _save_bytearray_data(self, obj, (const char*) view->buf,
|
|
view->len);
|
|
}
|
|
}
|
|
else {
|
|
/* Write data out-of-band */
|
|
const char next_buffer_op = NEXT_BUFFER;
|
|
if (_Pickler_Write(self, &next_buffer_op, 1) < 0) {
|
|
return -1;
|
|
}
|
|
if (view->readonly) {
|
|
const char readonly_buffer_op = READONLY_BUFFER;
|
|
if (_Pickler_Write(self, &readonly_buffer_op, 1) < 0) {
|
|
return -1;
|
|
}
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/* A copy of PyUnicode_EncodeRawUnicodeEscape() that also translates
|
|
backslash and newline characters to \uXXXX escapes. */
|
|
static PyObject *
|
|
raw_unicode_escape(PyObject *obj)
|
|
{
|
|
char *p;
|
|
Py_ssize_t i, size;
|
|
const void *data;
|
|
unsigned int kind;
|
|
_PyBytesWriter writer;
|
|
|
|
if (PyUnicode_READY(obj))
|
|
return NULL;
|
|
|
|
_PyBytesWriter_Init(&writer);
|
|
|
|
size = PyUnicode_GET_LENGTH(obj);
|
|
data = PyUnicode_DATA(obj);
|
|
kind = PyUnicode_KIND(obj);
|
|
|
|
p = _PyBytesWriter_Alloc(&writer, size);
|
|
if (p == NULL)
|
|
goto error;
|
|
writer.overallocate = 1;
|
|
|
|
for (i=0; i < size; i++) {
|
|
Py_UCS4 ch = PyUnicode_READ(kind, data, i);
|
|
/* Map 32-bit characters to '\Uxxxxxxxx' */
|
|
if (ch >= 0x10000) {
|
|
/* -1: subtract 1 preallocated byte */
|
|
p = _PyBytesWriter_Prepare(&writer, p, 10-1);
|
|
if (p == NULL)
|
|
goto error;
|
|
|
|
*p++ = '\\';
|
|
*p++ = 'U';
|
|
*p++ = Py_hexdigits[(ch >> 28) & 0xf];
|
|
*p++ = Py_hexdigits[(ch >> 24) & 0xf];
|
|
*p++ = Py_hexdigits[(ch >> 20) & 0xf];
|
|
*p++ = Py_hexdigits[(ch >> 16) & 0xf];
|
|
*p++ = Py_hexdigits[(ch >> 12) & 0xf];
|
|
*p++ = Py_hexdigits[(ch >> 8) & 0xf];
|
|
*p++ = Py_hexdigits[(ch >> 4) & 0xf];
|
|
*p++ = Py_hexdigits[ch & 15];
|
|
}
|
|
/* Map 16-bit characters, '\\' and '\n' to '\uxxxx' */
|
|
else if (ch >= 256 ||
|
|
ch == '\\' || ch == 0 || ch == '\n' || ch == '\r' ||
|
|
ch == 0x1a)
|
|
{
|
|
/* -1: subtract 1 preallocated byte */
|
|
p = _PyBytesWriter_Prepare(&writer, p, 6-1);
|
|
if (p == NULL)
|
|
goto error;
|
|
|
|
*p++ = '\\';
|
|
*p++ = 'u';
|
|
*p++ = Py_hexdigits[(ch >> 12) & 0xf];
|
|
*p++ = Py_hexdigits[(ch >> 8) & 0xf];
|
|
*p++ = Py_hexdigits[(ch >> 4) & 0xf];
|
|
*p++ = Py_hexdigits[ch & 15];
|
|
}
|
|
/* Copy everything else as-is */
|
|
else
|
|
*p++ = (char) ch;
|
|
}
|
|
|
|
return _PyBytesWriter_Finish(&writer, p);
|
|
|
|
error:
|
|
_PyBytesWriter_Dealloc(&writer);
|
|
return NULL;
|
|
}
|
|
|
|
static int
|
|
write_unicode_binary(PicklerObject *self, PyObject *obj)
|
|
{
|
|
char header[9];
|
|
Py_ssize_t len;
|
|
PyObject *encoded = NULL;
|
|
Py_ssize_t size;
|
|
const char *data;
|
|
|
|
if (PyUnicode_READY(obj))
|
|
return -1;
|
|
|
|
data = PyUnicode_AsUTF8AndSize(obj, &size);
|
|
if (data == NULL) {
|
|
/* Issue #8383: for strings with lone surrogates, fallback on the
|
|
"surrogatepass" error handler. */
|
|
PyErr_Clear();
|
|
encoded = PyUnicode_AsEncodedString(obj, "utf-8", "surrogatepass");
|
|
if (encoded == NULL)
|
|
return -1;
|
|
|
|
data = PyBytes_AS_STRING(encoded);
|
|
size = PyBytes_GET_SIZE(encoded);
|
|
}
|
|
|
|
assert(size >= 0);
|
|
if (size <= 0xff && self->proto >= 4) {
|
|
header[0] = SHORT_BINUNICODE;
|
|
header[1] = (unsigned char)(size & 0xff);
|
|
len = 2;
|
|
}
|
|
else if ((size_t)size <= 0xffffffffUL) {
|
|
header[0] = BINUNICODE;
|
|
header[1] = (unsigned char)(size & 0xff);
|
|
header[2] = (unsigned char)((size >> 8) & 0xff);
|
|
header[3] = (unsigned char)((size >> 16) & 0xff);
|
|
header[4] = (unsigned char)((size >> 24) & 0xff);
|
|
len = 5;
|
|
}
|
|
else if (self->proto >= 4) {
|
|
header[0] = BINUNICODE8;
|
|
_write_size64(header + 1, size);
|
|
len = 9;
|
|
}
|
|
else {
|
|
PyErr_SetString(PyExc_OverflowError,
|
|
"serializing a string larger than 4 GiB "
|
|
"requires pickle protocol 4 or higher");
|
|
Py_XDECREF(encoded);
|
|
return -1;
|
|
}
|
|
|
|
if (_Pickler_write_bytes(self, header, len, data, size, encoded) < 0) {
|
|
Py_XDECREF(encoded);
|
|
return -1;
|
|
}
|
|
Py_XDECREF(encoded);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
save_unicode(PicklerObject *self, PyObject *obj)
|
|
{
|
|
if (self->bin) {
|
|
if (write_unicode_binary(self, obj) < 0)
|
|
return -1;
|
|
}
|
|
else {
|
|
PyObject *encoded;
|
|
Py_ssize_t size;
|
|
const char unicode_op = UNICODE;
|
|
|
|
encoded = raw_unicode_escape(obj);
|
|
if (encoded == NULL)
|
|
return -1;
|
|
|
|
if (_Pickler_Write(self, &unicode_op, 1) < 0) {
|
|
Py_DECREF(encoded);
|
|
return -1;
|
|
}
|
|
|
|
size = PyBytes_GET_SIZE(encoded);
|
|
if (_Pickler_Write(self, PyBytes_AS_STRING(encoded), size) < 0) {
|
|
Py_DECREF(encoded);
|
|
return -1;
|
|
}
|
|
Py_DECREF(encoded);
|
|
|
|
if (_Pickler_Write(self, "\n", 1) < 0)
|
|
return -1;
|
|
}
|
|
if (memo_put(self, obj) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* A helper for save_tuple. Push the len elements in tuple t on the stack. */
|
|
static int
|
|
store_tuple_elements(PicklerObject *self, PyObject *t, Py_ssize_t len)
|
|
{
|
|
Py_ssize_t i;
|
|
|
|
assert(PyTuple_Size(t) == len);
|
|
|
|
for (i = 0; i < len; i++) {
|
|
PyObject *element = PyTuple_GET_ITEM(t, i);
|
|
|
|
if (element == NULL)
|
|
return -1;
|
|
if (save(self, element, 0) < 0)
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Tuples are ubiquitous in the pickle protocols, so many techniques are
|
|
* used across protocols to minimize the space needed to pickle them.
|
|
* Tuples are also the only builtin immutable type that can be recursive
|
|
* (a tuple can be reached from itself), and that requires some subtle
|
|
* magic so that it works in all cases. IOW, this is a long routine.
|
|
*/
|
|
static int
|
|
save_tuple(PicklerObject *self, PyObject *obj)
|
|
{
|
|
Py_ssize_t len, i;
|
|
|
|
const char mark_op = MARK;
|
|
const char tuple_op = TUPLE;
|
|
const char pop_op = POP;
|
|
const char pop_mark_op = POP_MARK;
|
|
const char len2opcode[] = {EMPTY_TUPLE, TUPLE1, TUPLE2, TUPLE3};
|
|
|
|
if ((len = PyTuple_Size(obj)) < 0)
|
|
return -1;
|
|
|
|
if (len == 0) {
|
|
char pdata[2];
|
|
|
|
if (self->proto) {
|
|
pdata[0] = EMPTY_TUPLE;
|
|
len = 1;
|
|
}
|
|
else {
|
|
pdata[0] = MARK;
|
|
pdata[1] = TUPLE;
|
|
len = 2;
|
|
}
|
|
if (_Pickler_Write(self, pdata, len) < 0)
|
|
return -1;
|
|
return 0;
|
|
}
|
|
|
|
/* The tuple isn't in the memo now. If it shows up there after
|
|
* saving the tuple elements, the tuple must be recursive, in
|
|
* which case we'll pop everything we put on the stack, and fetch
|
|
* its value from the memo.
|
|
*/
|
|
if (len <= 3 && self->proto >= 2) {
|
|
/* Use TUPLE{1,2,3} opcodes. */
|
|
if (store_tuple_elements(self, obj, len) < 0)
|
|
return -1;
|
|
|
|
if (PyMemoTable_Get(self->memo, obj)) {
|
|
/* pop the len elements */
|
|
for (i = 0; i < len; i++)
|
|
if (_Pickler_Write(self, &pop_op, 1) < 0)
|
|
return -1;
|
|
/* fetch from memo */
|
|
if (memo_get(self, obj) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
else { /* Not recursive. */
|
|
if (_Pickler_Write(self, len2opcode + len, 1) < 0)
|
|
return -1;
|
|
}
|
|
goto memoize;
|
|
}
|
|
|
|
/* proto < 2 and len > 0, or proto >= 2 and len > 3.
|
|
* Generate MARK e1 e2 ... TUPLE
|
|
*/
|
|
if (_Pickler_Write(self, &mark_op, 1) < 0)
|
|
return -1;
|
|
|
|
if (store_tuple_elements(self, obj, len) < 0)
|
|
return -1;
|
|
|
|
if (PyMemoTable_Get(self->memo, obj)) {
|
|
/* pop the stack stuff we pushed */
|
|
if (self->bin) {
|
|
if (_Pickler_Write(self, &pop_mark_op, 1) < 0)
|
|
return -1;
|
|
}
|
|
else {
|
|
/* Note that we pop one more than len, to remove
|
|
* the MARK too.
|
|
*/
|
|
for (i = 0; i <= len; i++)
|
|
if (_Pickler_Write(self, &pop_op, 1) < 0)
|
|
return -1;
|
|
}
|
|
/* fetch from memo */
|
|
if (memo_get(self, obj) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
else { /* Not recursive. */
|
|
if (_Pickler_Write(self, &tuple_op, 1) < 0)
|
|
return -1;
|
|
}
|
|
|
|
memoize:
|
|
if (memo_put(self, obj) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* iter is an iterator giving items, and we batch up chunks of
|
|
* MARK item item ... item APPENDS
|
|
* opcode sequences. Calling code should have arranged to first create an
|
|
* empty list, or list-like object, for the APPENDS to operate on.
|
|
* Returns 0 on success, <0 on error.
|
|
*/
|
|
static int
|
|
batch_list(PicklerObject *self, PyObject *iter)
|
|
{
|
|
PyObject *obj = NULL;
|
|
PyObject *firstitem = NULL;
|
|
int i, n;
|
|
|
|
const char mark_op = MARK;
|
|
const char append_op = APPEND;
|
|
const char appends_op = APPENDS;
|
|
|
|
assert(iter != NULL);
|
|
|
|
/* XXX: I think this function could be made faster by avoiding the
|
|
iterator interface and fetching objects directly from list using
|
|
PyList_GET_ITEM.
|
|
*/
|
|
|
|
if (self->proto == 0) {
|
|
/* APPENDS isn't available; do one at a time. */
|
|
for (;;) {
|
|
obj = PyIter_Next(iter);
|
|
if (obj == NULL) {
|
|
if (PyErr_Occurred())
|
|
return -1;
|
|
break;
|
|
}
|
|
i = save(self, obj, 0);
|
|
Py_DECREF(obj);
|
|
if (i < 0)
|
|
return -1;
|
|
if (_Pickler_Write(self, &append_op, 1) < 0)
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/* proto > 0: write in batches of BATCHSIZE. */
|
|
do {
|
|
/* Get first item */
|
|
firstitem = PyIter_Next(iter);
|
|
if (firstitem == NULL) {
|
|
if (PyErr_Occurred())
|
|
goto error;
|
|
|
|
/* nothing more to add */
|
|
break;
|
|
}
|
|
|
|
/* Try to get a second item */
|
|
obj = PyIter_Next(iter);
|
|
if (obj == NULL) {
|
|
if (PyErr_Occurred())
|
|
goto error;
|
|
|
|
/* Only one item to write */
|
|
if (save(self, firstitem, 0) < 0)
|
|
goto error;
|
|
if (_Pickler_Write(self, &append_op, 1) < 0)
|
|
goto error;
|
|
Py_CLEAR(firstitem);
|
|
break;
|
|
}
|
|
|
|
/* More than one item to write */
|
|
|
|
/* Pump out MARK, items, APPENDS. */
|
|
if (_Pickler_Write(self, &mark_op, 1) < 0)
|
|
goto error;
|
|
|
|
if (save(self, firstitem, 0) < 0)
|
|
goto error;
|
|
Py_CLEAR(firstitem);
|
|
n = 1;
|
|
|
|
/* Fetch and save up to BATCHSIZE items */
|
|
while (obj) {
|
|
if (save(self, obj, 0) < 0)
|
|
goto error;
|
|
Py_CLEAR(obj);
|
|
n += 1;
|
|
|
|
if (n == BATCHSIZE)
|
|
break;
|
|
|
|
obj = PyIter_Next(iter);
|
|
if (obj == NULL) {
|
|
if (PyErr_Occurred())
|
|
goto error;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (_Pickler_Write(self, &appends_op, 1) < 0)
|
|
goto error;
|
|
|
|
} while (n == BATCHSIZE);
|
|
return 0;
|
|
|
|
error:
|
|
Py_XDECREF(firstitem);
|
|
Py_XDECREF(obj);
|
|
return -1;
|
|
}
|
|
|
|
/* This is a variant of batch_list() above, specialized for lists (with no
|
|
* support for list subclasses). Like batch_list(), we batch up chunks of
|
|
* MARK item item ... item APPENDS
|
|
* opcode sequences. Calling code should have arranged to first create an
|
|
* empty list, or list-like object, for the APPENDS to operate on.
|
|
* Returns 0 on success, -1 on error.
|
|
*
|
|
* This version is considerably faster than batch_list(), if less general.
|
|
*
|
|
* Note that this only works for protocols > 0.
|
|
*/
|
|
static int
|
|
batch_list_exact(PicklerObject *self, PyObject *obj)
|
|
{
|
|
PyObject *item = NULL;
|
|
Py_ssize_t this_batch, total;
|
|
|
|
const char append_op = APPEND;
|
|
const char appends_op = APPENDS;
|
|
const char mark_op = MARK;
|
|
|
|
assert(obj != NULL);
|
|
assert(self->proto > 0);
|
|
assert(PyList_CheckExact(obj));
|
|
|
|
if (PyList_GET_SIZE(obj) == 1) {
|
|
item = PyList_GET_ITEM(obj, 0);
|
|
if (save(self, item, 0) < 0)
|
|
return -1;
|
|
if (_Pickler_Write(self, &append_op, 1) < 0)
|
|
return -1;
|
|
return 0;
|
|
}
|
|
|
|
/* Write in batches of BATCHSIZE. */
|
|
total = 0;
|
|
do {
|
|
this_batch = 0;
|
|
if (_Pickler_Write(self, &mark_op, 1) < 0)
|
|
return -1;
|
|
while (total < PyList_GET_SIZE(obj)) {
|
|
item = PyList_GET_ITEM(obj, total);
|
|
if (save(self, item, 0) < 0)
|
|
return -1;
|
|
total++;
|
|
if (++this_batch == BATCHSIZE)
|
|
break;
|
|
}
|
|
if (_Pickler_Write(self, &appends_op, 1) < 0)
|
|
return -1;
|
|
|
|
} while (total < PyList_GET_SIZE(obj));
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
save_list(PicklerObject *self, PyObject *obj)
|
|
{
|
|
char header[3];
|
|
Py_ssize_t len;
|
|
int status = 0;
|
|
|
|
if (self->fast && !fast_save_enter(self, obj))
|
|
goto error;
|
|
|
|
/* Create an empty list. */
|
|
if (self->bin) {
|
|
header[0] = EMPTY_LIST;
|
|
len = 1;
|
|
}
|
|
else {
|
|
header[0] = MARK;
|
|
header[1] = LIST;
|
|
len = 2;
|
|
}
|
|
|
|
if (_Pickler_Write(self, header, len) < 0)
|
|
goto error;
|
|
|
|
/* Get list length, and bow out early if empty. */
|
|
if ((len = PyList_Size(obj)) < 0)
|
|
goto error;
|
|
|
|
if (memo_put(self, obj) < 0)
|
|
goto error;
|
|
|
|
if (len != 0) {
|
|
/* Materialize the list elements. */
|
|
if (PyList_CheckExact(obj) && self->proto > 0) {
|
|
if (Py_EnterRecursiveCall(" while pickling an object"))
|
|
goto error;
|
|
status = batch_list_exact(self, obj);
|
|
Py_LeaveRecursiveCall();
|
|
} else {
|
|
PyObject *iter = PyObject_GetIter(obj);
|
|
if (iter == NULL)
|
|
goto error;
|
|
|
|
if (Py_EnterRecursiveCall(" while pickling an object")) {
|
|
Py_DECREF(iter);
|
|
goto error;
|
|
}
|
|
status = batch_list(self, iter);
|
|
Py_LeaveRecursiveCall();
|
|
Py_DECREF(iter);
|
|
}
|
|
}
|
|
if (0) {
|
|
error:
|
|
status = -1;
|
|
}
|
|
|
|
if (self->fast && !fast_save_leave(self, obj))
|
|
status = -1;
|
|
|
|
return status;
|
|
}
|
|
|
|
/* iter is an iterator giving (key, value) pairs, and we batch up chunks of
|
|
* MARK key value ... key value SETITEMS
|
|
* opcode sequences. Calling code should have arranged to first create an
|
|
* empty dict, or dict-like object, for the SETITEMS to operate on.
|
|
* Returns 0 on success, <0 on error.
|
|
*
|
|
* This is very much like batch_list(). The difference between saving
|
|
* elements directly, and picking apart two-tuples, is so long-winded at
|
|
* the C level, though, that attempts to combine these routines were too
|
|
* ugly to bear.
|
|
*/
|
|
static int
|
|
batch_dict(PicklerObject *self, PyObject *iter)
|
|
{
|
|
PyObject *obj = NULL;
|
|
PyObject *firstitem = NULL;
|
|
int i, n;
|
|
|
|
const char mark_op = MARK;
|
|
const char setitem_op = SETITEM;
|
|
const char setitems_op = SETITEMS;
|
|
|
|
assert(iter != NULL);
|
|
|
|
if (self->proto == 0) {
|
|
/* SETITEMS isn't available; do one at a time. */
|
|
for (;;) {
|
|
obj = PyIter_Next(iter);
|
|
if (obj == NULL) {
|
|
if (PyErr_Occurred())
|
|
return -1;
|
|
break;
|
|
}
|
|
if (!PyTuple_Check(obj) || PyTuple_Size(obj) != 2) {
|
|
PyErr_SetString(PyExc_TypeError, "dict items "
|
|
"iterator must return 2-tuples");
|
|
return -1;
|
|
}
|
|
i = save(self, PyTuple_GET_ITEM(obj, 0), 0);
|
|
if (i >= 0)
|
|
i = save(self, PyTuple_GET_ITEM(obj, 1), 0);
|
|
Py_DECREF(obj);
|
|
if (i < 0)
|
|
return -1;
|
|
if (_Pickler_Write(self, &setitem_op, 1) < 0)
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/* proto > 0: write in batches of BATCHSIZE. */
|
|
do {
|
|
/* Get first item */
|
|
firstitem = PyIter_Next(iter);
|
|
if (firstitem == NULL) {
|
|
if (PyErr_Occurred())
|
|
goto error;
|
|
|
|
/* nothing more to add */
|
|
break;
|
|
}
|
|
if (!PyTuple_Check(firstitem) || PyTuple_Size(firstitem) != 2) {
|
|
PyErr_SetString(PyExc_TypeError, "dict items "
|
|
"iterator must return 2-tuples");
|
|
goto error;
|
|
}
|
|
|
|
/* Try to get a second item */
|
|
obj = PyIter_Next(iter);
|
|
if (obj == NULL) {
|
|
if (PyErr_Occurred())
|
|
goto error;
|
|
|
|
/* Only one item to write */
|
|
if (save(self, PyTuple_GET_ITEM(firstitem, 0), 0) < 0)
|
|
goto error;
|
|
if (save(self, PyTuple_GET_ITEM(firstitem, 1), 0) < 0)
|
|
goto error;
|
|
if (_Pickler_Write(self, &setitem_op, 1) < 0)
|
|
goto error;
|
|
Py_CLEAR(firstitem);
|
|
break;
|
|
}
|
|
|
|
/* More than one item to write */
|
|
|
|
/* Pump out MARK, items, SETITEMS. */
|
|
if (_Pickler_Write(self, &mark_op, 1) < 0)
|
|
goto error;
|
|
|
|
if (save(self, PyTuple_GET_ITEM(firstitem, 0), 0) < 0)
|
|
goto error;
|
|
if (save(self, PyTuple_GET_ITEM(firstitem, 1), 0) < 0)
|
|
goto error;
|
|
Py_CLEAR(firstitem);
|
|
n = 1;
|
|
|
|
/* Fetch and save up to BATCHSIZE items */
|
|
while (obj) {
|
|
if (!PyTuple_Check(obj) || PyTuple_Size(obj) != 2) {
|
|
PyErr_SetString(PyExc_TypeError, "dict items "
|
|
"iterator must return 2-tuples");
|
|
goto error;
|
|
}
|
|
if (save(self, PyTuple_GET_ITEM(obj, 0), 0) < 0 ||
|
|
save(self, PyTuple_GET_ITEM(obj, 1), 0) < 0)
|
|
goto error;
|
|
Py_CLEAR(obj);
|
|
n += 1;
|
|
|
|
if (n == BATCHSIZE)
|
|
break;
|
|
|
|
obj = PyIter_Next(iter);
|
|
if (obj == NULL) {
|
|
if (PyErr_Occurred())
|
|
goto error;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (_Pickler_Write(self, &setitems_op, 1) < 0)
|
|
goto error;
|
|
|
|
} while (n == BATCHSIZE);
|
|
return 0;
|
|
|
|
error:
|
|
Py_XDECREF(firstitem);
|
|
Py_XDECREF(obj);
|
|
return -1;
|
|
}
|
|
|
|
/* This is a variant of batch_dict() above that specializes for dicts, with no
|
|
* support for dict subclasses. Like batch_dict(), we batch up chunks of
|
|
* MARK key value ... key value SETITEMS
|
|
* opcode sequences. Calling code should have arranged to first create an
|
|
* empty dict, or dict-like object, for the SETITEMS to operate on.
|
|
* Returns 0 on success, -1 on error.
|
|
*
|
|
* Note that this currently doesn't work for protocol 0.
|
|
*/
|
|
static int
|
|
batch_dict_exact(PicklerObject *self, PyObject *obj)
|
|
{
|
|
PyObject *key = NULL, *value = NULL;
|
|
int i;
|
|
Py_ssize_t dict_size, ppos = 0;
|
|
|
|
const char mark_op = MARK;
|
|
const char setitem_op = SETITEM;
|
|
const char setitems_op = SETITEMS;
|
|
|
|
assert(obj != NULL && PyDict_CheckExact(obj));
|
|
assert(self->proto > 0);
|
|
|
|
dict_size = PyDict_GET_SIZE(obj);
|
|
|
|
/* Special-case len(d) == 1 to save space. */
|
|
if (dict_size == 1) {
|
|
PyDict_Next(obj, &ppos, &key, &value);
|
|
if (save(self, key, 0) < 0)
|
|
return -1;
|
|
if (save(self, value, 0) < 0)
|
|
return -1;
|
|
if (_Pickler_Write(self, &setitem_op, 1) < 0)
|
|
return -1;
|
|
return 0;
|
|
}
|
|
|
|
/* Write in batches of BATCHSIZE. */
|
|
do {
|
|
i = 0;
|
|
if (_Pickler_Write(self, &mark_op, 1) < 0)
|
|
return -1;
|
|
while (PyDict_Next(obj, &ppos, &key, &value)) {
|
|
if (save(self, key, 0) < 0)
|
|
return -1;
|
|
if (save(self, value, 0) < 0)
|
|
return -1;
|
|
if (++i == BATCHSIZE)
|
|
break;
|
|
}
|
|
if (_Pickler_Write(self, &setitems_op, 1) < 0)
|
|
return -1;
|
|
if (PyDict_GET_SIZE(obj) != dict_size) {
|
|
PyErr_Format(
|
|
PyExc_RuntimeError,
|
|
"dictionary changed size during iteration");
|
|
return -1;
|
|
}
|
|
|
|
} while (i == BATCHSIZE);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
save_dict(PicklerObject *self, PyObject *obj)
|
|
{
|
|
PyObject *items, *iter;
|
|
char header[3];
|
|
Py_ssize_t len;
|
|
int status = 0;
|
|
assert(PyDict_Check(obj));
|
|
|
|
if (self->fast && !fast_save_enter(self, obj))
|
|
goto error;
|
|
|
|
/* Create an empty dict. */
|
|
if (self->bin) {
|
|
header[0] = EMPTY_DICT;
|
|
len = 1;
|
|
}
|
|
else {
|
|
header[0] = MARK;
|
|
header[1] = DICT;
|
|
len = 2;
|
|
}
|
|
|
|
if (_Pickler_Write(self, header, len) < 0)
|
|
goto error;
|
|
|
|
if (memo_put(self, obj) < 0)
|
|
goto error;
|
|
|
|
if (PyDict_GET_SIZE(obj)) {
|
|
/* Save the dict items. */
|
|
if (PyDict_CheckExact(obj) && self->proto > 0) {
|
|
/* We can take certain shortcuts if we know this is a dict and
|
|
not a dict subclass. */
|
|
if (Py_EnterRecursiveCall(" while pickling an object"))
|
|
goto error;
|
|
status = batch_dict_exact(self, obj);
|
|
Py_LeaveRecursiveCall();
|
|
} else {
|
|
_Py_IDENTIFIER(items);
|
|
|
|
items = _PyObject_CallMethodIdNoArgs(obj, &PyId_items);
|
|
if (items == NULL)
|
|
goto error;
|
|
iter = PyObject_GetIter(items);
|
|
Py_DECREF(items);
|
|
if (iter == NULL)
|
|
goto error;
|
|
if (Py_EnterRecursiveCall(" while pickling an object")) {
|
|
Py_DECREF(iter);
|
|
goto error;
|
|
}
|
|
status = batch_dict(self, iter);
|
|
Py_LeaveRecursiveCall();
|
|
Py_DECREF(iter);
|
|
}
|
|
}
|
|
|
|
if (0) {
|
|
error:
|
|
status = -1;
|
|
}
|
|
|
|
if (self->fast && !fast_save_leave(self, obj))
|
|
status = -1;
|
|
|
|
return status;
|
|
}
|
|
|
|
static int
|
|
save_set(PicklerObject *self, PyObject *obj)
|
|
{
|
|
PyObject *item;
|
|
int i;
|
|
Py_ssize_t set_size, ppos = 0;
|
|
Py_hash_t hash;
|
|
|
|
const char empty_set_op = EMPTY_SET;
|
|
const char mark_op = MARK;
|
|
const char additems_op = ADDITEMS;
|
|
|
|
if (self->proto < 4) {
|
|
PyObject *items;
|
|
PyObject *reduce_value;
|
|
int status;
|
|
|
|
items = PySequence_List(obj);
|
|
if (items == NULL) {
|
|
return -1;
|
|
}
|
|
reduce_value = Py_BuildValue("(O(O))", (PyObject*)&PySet_Type, items);
|
|
Py_DECREF(items);
|
|
if (reduce_value == NULL) {
|
|
return -1;
|
|
}
|
|
/* save_reduce() will memoize the object automatically. */
|
|
status = save_reduce(self, reduce_value, obj);
|
|
Py_DECREF(reduce_value);
|
|
return status;
|
|
}
|
|
|
|
if (_Pickler_Write(self, &empty_set_op, 1) < 0)
|
|
return -1;
|
|
|
|
if (memo_put(self, obj) < 0)
|
|
return -1;
|
|
|
|
set_size = PySet_GET_SIZE(obj);
|
|
if (set_size == 0)
|
|
return 0; /* nothing to do */
|
|
|
|
/* Write in batches of BATCHSIZE. */
|
|
do {
|
|
i = 0;
|
|
if (_Pickler_Write(self, &mark_op, 1) < 0)
|
|
return -1;
|
|
while (_PySet_NextEntry(obj, &ppos, &item, &hash)) {
|
|
if (save(self, item, 0) < 0)
|
|
return -1;
|
|
if (++i == BATCHSIZE)
|
|
break;
|
|
}
|
|
if (_Pickler_Write(self, &additems_op, 1) < 0)
|
|
return -1;
|
|
if (PySet_GET_SIZE(obj) != set_size) {
|
|
PyErr_Format(
|
|
PyExc_RuntimeError,
|
|
"set changed size during iteration");
|
|
return -1;
|
|
}
|
|
} while (i == BATCHSIZE);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
save_frozenset(PicklerObject *self, PyObject *obj)
|
|
{
|
|
PyObject *iter;
|
|
|
|
const char mark_op = MARK;
|
|
const char frozenset_op = FROZENSET;
|
|
|
|
if (self->fast && !fast_save_enter(self, obj))
|
|
return -1;
|
|
|
|
if (self->proto < 4) {
|
|
PyObject *items;
|
|
PyObject *reduce_value;
|
|
int status;
|
|
|
|
items = PySequence_List(obj);
|
|
if (items == NULL) {
|
|
return -1;
|
|
}
|
|
reduce_value = Py_BuildValue("(O(O))", (PyObject*)&PyFrozenSet_Type,
|
|
items);
|
|
Py_DECREF(items);
|
|
if (reduce_value == NULL) {
|
|
return -1;
|
|
}
|
|
/* save_reduce() will memoize the object automatically. */
|
|
status = save_reduce(self, reduce_value, obj);
|
|
Py_DECREF(reduce_value);
|
|
return status;
|
|
}
|
|
|
|
if (_Pickler_Write(self, &mark_op, 1) < 0)
|
|
return -1;
|
|
|
|
iter = PyObject_GetIter(obj);
|
|
if (iter == NULL) {
|
|
return -1;
|
|
}
|
|
for (;;) {
|
|
PyObject *item;
|
|
|
|
item = PyIter_Next(iter);
|
|
if (item == NULL) {
|
|
if (PyErr_Occurred()) {
|
|
Py_DECREF(iter);
|
|
return -1;
|
|
}
|
|
break;
|
|
}
|
|
if (save(self, item, 0) < 0) {
|
|
Py_DECREF(item);
|
|
Py_DECREF(iter);
|
|
return -1;
|
|
}
|
|
Py_DECREF(item);
|
|
}
|
|
Py_DECREF(iter);
|
|
|
|
/* If the object is already in the memo, this means it is
|
|
recursive. In this case, throw away everything we put on the
|
|
stack, and fetch the object back from the memo. */
|
|
if (PyMemoTable_Get(self->memo, obj)) {
|
|
const char pop_mark_op = POP_MARK;
|
|
|
|
if (_Pickler_Write(self, &pop_mark_op, 1) < 0)
|
|
return -1;
|
|
if (memo_get(self, obj) < 0)
|
|
return -1;
|
|
return 0;
|
|
}
|
|
|
|
if (_Pickler_Write(self, &frozenset_op, 1) < 0)
|
|
return -1;
|
|
if (memo_put(self, obj) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
fix_imports(PyObject **module_name, PyObject **global_name)
|
|
{
|
|
PyObject *key;
|
|
PyObject *item;
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
|
|
key = PyTuple_Pack(2, *module_name, *global_name);
|
|
if (key == NULL)
|
|
return -1;
|
|
item = PyDict_GetItemWithError(st->name_mapping_3to2, key);
|
|
Py_DECREF(key);
|
|
if (item) {
|
|
PyObject *fixed_module_name;
|
|
PyObject *fixed_global_name;
|
|
|
|
if (!PyTuple_Check(item) || PyTuple_GET_SIZE(item) != 2) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"_compat_pickle.REVERSE_NAME_MAPPING values "
|
|
"should be 2-tuples, not %.200s",
|
|
Py_TYPE(item)->tp_name);
|
|
return -1;
|
|
}
|
|
fixed_module_name = PyTuple_GET_ITEM(item, 0);
|
|
fixed_global_name = PyTuple_GET_ITEM(item, 1);
|
|
if (!PyUnicode_Check(fixed_module_name) ||
|
|
!PyUnicode_Check(fixed_global_name)) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"_compat_pickle.REVERSE_NAME_MAPPING values "
|
|
"should be pairs of str, not (%.200s, %.200s)",
|
|
Py_TYPE(fixed_module_name)->tp_name,
|
|
Py_TYPE(fixed_global_name)->tp_name);
|
|
return -1;
|
|
}
|
|
|
|
Py_CLEAR(*module_name);
|
|
Py_CLEAR(*global_name);
|
|
Py_INCREF(fixed_module_name);
|
|
Py_INCREF(fixed_global_name);
|
|
*module_name = fixed_module_name;
|
|
*global_name = fixed_global_name;
|
|
return 0;
|
|
}
|
|
else if (PyErr_Occurred()) {
|
|
return -1;
|
|
}
|
|
|
|
item = PyDict_GetItemWithError(st->import_mapping_3to2, *module_name);
|
|
if (item) {
|
|
if (!PyUnicode_Check(item)) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"_compat_pickle.REVERSE_IMPORT_MAPPING values "
|
|
"should be strings, not %.200s",
|
|
Py_TYPE(item)->tp_name);
|
|
return -1;
|
|
}
|
|
Py_INCREF(item);
|
|
Py_XSETREF(*module_name, item);
|
|
}
|
|
else if (PyErr_Occurred()) {
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
save_global(PicklerObject *self, PyObject *obj, PyObject *name)
|
|
{
|
|
PyObject *global_name = NULL;
|
|
PyObject *module_name = NULL;
|
|
PyObject *module = NULL;
|
|
PyObject *parent = NULL;
|
|
PyObject *dotted_path = NULL;
|
|
PyObject *lastname = NULL;
|
|
PyObject *cls;
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
int status = 0;
|
|
_Py_IDENTIFIER(__name__);
|
|
_Py_IDENTIFIER(__qualname__);
|
|
|
|
const char global_op = GLOBAL;
|
|
|
|
if (name) {
|
|
Py_INCREF(name);
|
|
global_name = name;
|
|
}
|
|
else {
|
|
if (_PyObject_LookupAttrId(obj, &PyId___qualname__, &global_name) < 0)
|
|
goto error;
|
|
if (global_name == NULL) {
|
|
global_name = _PyObject_GetAttrId(obj, &PyId___name__);
|
|
if (global_name == NULL)
|
|
goto error;
|
|
}
|
|
}
|
|
|
|
dotted_path = get_dotted_path(module, global_name);
|
|
if (dotted_path == NULL)
|
|
goto error;
|
|
module_name = whichmodule(obj, dotted_path);
|
|
if (module_name == NULL)
|
|
goto error;
|
|
|
|
/* XXX: Change to use the import C API directly with level=0 to disallow
|
|
relative imports.
|
|
|
|
XXX: PyImport_ImportModuleLevel could be used. However, this bypasses
|
|
builtins.__import__. Therefore, _pickle, unlike pickle.py, will ignore
|
|
custom import functions (IMHO, this would be a nice security
|
|
feature). The import C API would need to be extended to support the
|
|
extra parameters of __import__ to fix that. */
|
|
module = PyImport_Import(module_name);
|
|
if (module == NULL) {
|
|
PyErr_Format(st->PicklingError,
|
|
"Can't pickle %R: import of module %R failed",
|
|
obj, module_name);
|
|
goto error;
|
|
}
|
|
lastname = PyList_GET_ITEM(dotted_path, PyList_GET_SIZE(dotted_path)-1);
|
|
Py_INCREF(lastname);
|
|
cls = get_deep_attribute(module, dotted_path, &parent);
|
|
Py_CLEAR(dotted_path);
|
|
if (cls == NULL) {
|
|
PyErr_Format(st->PicklingError,
|
|
"Can't pickle %R: attribute lookup %S on %S failed",
|
|
obj, global_name, module_name);
|
|
goto error;
|
|
}
|
|
if (cls != obj) {
|
|
Py_DECREF(cls);
|
|
PyErr_Format(st->PicklingError,
|
|
"Can't pickle %R: it's not the same object as %S.%S",
|
|
obj, module_name, global_name);
|
|
goto error;
|
|
}
|
|
Py_DECREF(cls);
|
|
|
|
if (self->proto >= 2) {
|
|
/* See whether this is in the extension registry, and if
|
|
* so generate an EXT opcode.
|
|
*/
|
|
PyObject *extension_key;
|
|
PyObject *code_obj; /* extension code as Python object */
|
|
long code; /* extension code as C value */
|
|
char pdata[5];
|
|
Py_ssize_t n;
|
|
|
|
extension_key = PyTuple_Pack(2, module_name, global_name);
|
|
if (extension_key == NULL) {
|
|
goto error;
|
|
}
|
|
code_obj = PyDict_GetItemWithError(st->extension_registry,
|
|
extension_key);
|
|
Py_DECREF(extension_key);
|
|
/* The object is not registered in the extension registry.
|
|
This is the most likely code path. */
|
|
if (code_obj == NULL) {
|
|
if (PyErr_Occurred()) {
|
|
goto error;
|
|
}
|
|
goto gen_global;
|
|
}
|
|
|
|
/* XXX: pickle.py doesn't check neither the type, nor the range
|
|
of the value returned by the extension_registry. It should for
|
|
consistency. */
|
|
|
|
/* Verify code_obj has the right type and value. */
|
|
if (!PyLong_Check(code_obj)) {
|
|
PyErr_Format(st->PicklingError,
|
|
"Can't pickle %R: extension code %R isn't an integer",
|
|
obj, code_obj);
|
|
goto error;
|
|
}
|
|
code = PyLong_AS_LONG(code_obj);
|
|
if (code <= 0 || code > 0x7fffffffL) {
|
|
if (!PyErr_Occurred())
|
|
PyErr_Format(st->PicklingError, "Can't pickle %R: extension "
|
|
"code %ld is out of range", obj, code);
|
|
goto error;
|
|
}
|
|
|
|
/* Generate an EXT opcode. */
|
|
if (code <= 0xff) {
|
|
pdata[0] = EXT1;
|
|
pdata[1] = (unsigned char)code;
|
|
n = 2;
|
|
}
|
|
else if (code <= 0xffff) {
|
|
pdata[0] = EXT2;
|
|
pdata[1] = (unsigned char)(code & 0xff);
|
|
pdata[2] = (unsigned char)((code >> 8) & 0xff);
|
|
n = 3;
|
|
}
|
|
else {
|
|
pdata[0] = EXT4;
|
|
pdata[1] = (unsigned char)(code & 0xff);
|
|
pdata[2] = (unsigned char)((code >> 8) & 0xff);
|
|
pdata[3] = (unsigned char)((code >> 16) & 0xff);
|
|
pdata[4] = (unsigned char)((code >> 24) & 0xff);
|
|
n = 5;
|
|
}
|
|
|
|
if (_Pickler_Write(self, pdata, n) < 0)
|
|
goto error;
|
|
}
|
|
else {
|
|
gen_global:
|
|
if (parent == module) {
|
|
Py_INCREF(lastname);
|
|
Py_DECREF(global_name);
|
|
global_name = lastname;
|
|
}
|
|
if (self->proto >= 4) {
|
|
const char stack_global_op = STACK_GLOBAL;
|
|
|
|
if (save(self, module_name, 0) < 0)
|
|
goto error;
|
|
if (save(self, global_name, 0) < 0)
|
|
goto error;
|
|
|
|
if (_Pickler_Write(self, &stack_global_op, 1) < 0)
|
|
goto error;
|
|
}
|
|
else if (parent != module) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyObject *reduce_value = Py_BuildValue("(O(OO))",
|
|
st->getattr, parent, lastname);
|
|
if (reduce_value == NULL)
|
|
goto error;
|
|
status = save_reduce(self, reduce_value, NULL);
|
|
Py_DECREF(reduce_value);
|
|
if (status < 0)
|
|
goto error;
|
|
}
|
|
else {
|
|
/* Generate a normal global opcode if we are using a pickle
|
|
protocol < 4, or if the object is not registered in the
|
|
extension registry. */
|
|
PyObject *encoded;
|
|
PyObject *(*unicode_encoder)(PyObject *);
|
|
|
|
if (_Pickler_Write(self, &global_op, 1) < 0)
|
|
goto error;
|
|
|
|
/* For protocol < 3 and if the user didn't request against doing
|
|
so, we convert module names to the old 2.x module names. */
|
|
if (self->proto < 3 && self->fix_imports) {
|
|
if (fix_imports(&module_name, &global_name) < 0) {
|
|
goto error;
|
|
}
|
|
}
|
|
|
|
/* Since Python 3.0 now supports non-ASCII identifiers, we encode
|
|
both the module name and the global name using UTF-8. We do so
|
|
only when we are using the pickle protocol newer than version
|
|
3. This is to ensure compatibility with older Unpickler running
|
|
on Python 2.x. */
|
|
if (self->proto == 3) {
|
|
unicode_encoder = PyUnicode_AsUTF8String;
|
|
}
|
|
else {
|
|
unicode_encoder = PyUnicode_AsASCIIString;
|
|
}
|
|
encoded = unicode_encoder(module_name);
|
|
if (encoded == NULL) {
|
|
if (PyErr_ExceptionMatches(PyExc_UnicodeEncodeError))
|
|
PyErr_Format(st->PicklingError,
|
|
"can't pickle module identifier '%S' using "
|
|
"pickle protocol %i",
|
|
module_name, self->proto);
|
|
goto error;
|
|
}
|
|
if (_Pickler_Write(self, PyBytes_AS_STRING(encoded),
|
|
PyBytes_GET_SIZE(encoded)) < 0) {
|
|
Py_DECREF(encoded);
|
|
goto error;
|
|
}
|
|
Py_DECREF(encoded);
|
|
if(_Pickler_Write(self, "\n", 1) < 0)
|
|
goto error;
|
|
|
|
/* Save the name of the module. */
|
|
encoded = unicode_encoder(global_name);
|
|
if (encoded == NULL) {
|
|
if (PyErr_ExceptionMatches(PyExc_UnicodeEncodeError))
|
|
PyErr_Format(st->PicklingError,
|
|
"can't pickle global identifier '%S' using "
|
|
"pickle protocol %i",
|
|
global_name, self->proto);
|
|
goto error;
|
|
}
|
|
if (_Pickler_Write(self, PyBytes_AS_STRING(encoded),
|
|
PyBytes_GET_SIZE(encoded)) < 0) {
|
|
Py_DECREF(encoded);
|
|
goto error;
|
|
}
|
|
Py_DECREF(encoded);
|
|
if (_Pickler_Write(self, "\n", 1) < 0)
|
|
goto error;
|
|
}
|
|
/* Memoize the object. */
|
|
if (memo_put(self, obj) < 0)
|
|
goto error;
|
|
}
|
|
|
|
if (0) {
|
|
error:
|
|
status = -1;
|
|
}
|
|
Py_XDECREF(module_name);
|
|
Py_XDECREF(global_name);
|
|
Py_XDECREF(module);
|
|
Py_XDECREF(parent);
|
|
Py_XDECREF(dotted_path);
|
|
Py_XDECREF(lastname);
|
|
|
|
return status;
|
|
}
|
|
|
|
static int
|
|
save_singleton_type(PicklerObject *self, PyObject *obj, PyObject *singleton)
|
|
{
|
|
PyObject *reduce_value;
|
|
int status;
|
|
|
|
reduce_value = Py_BuildValue("O(O)", &PyType_Type, singleton);
|
|
if (reduce_value == NULL) {
|
|
return -1;
|
|
}
|
|
status = save_reduce(self, reduce_value, obj);
|
|
Py_DECREF(reduce_value);
|
|
return status;
|
|
}
|
|
|
|
static int
|
|
save_type(PicklerObject *self, PyObject *obj)
|
|
{
|
|
if (obj == (PyObject *)&_PyNone_Type) {
|
|
return save_singleton_type(self, obj, Py_None);
|
|
}
|
|
else if (obj == (PyObject *)&PyEllipsis_Type) {
|
|
return save_singleton_type(self, obj, Py_Ellipsis);
|
|
}
|
|
else if (obj == (PyObject *)&_PyNotImplemented_Type) {
|
|
return save_singleton_type(self, obj, Py_NotImplemented);
|
|
}
|
|
return save_global(self, obj, NULL);
|
|
}
|
|
|
|
static int
|
|
save_pers(PicklerObject *self, PyObject *obj)
|
|
{
|
|
PyObject *pid = NULL;
|
|
int status = 0;
|
|
|
|
const char persid_op = PERSID;
|
|
const char binpersid_op = BINPERSID;
|
|
|
|
pid = call_method(self->pers_func, self->pers_func_self, obj);
|
|
if (pid == NULL)
|
|
return -1;
|
|
|
|
if (pid != Py_None) {
|
|
if (self->bin) {
|
|
if (save(self, pid, 1) < 0 ||
|
|
_Pickler_Write(self, &binpersid_op, 1) < 0)
|
|
goto error;
|
|
}
|
|
else {
|
|
PyObject *pid_str;
|
|
|
|
pid_str = PyObject_Str(pid);
|
|
if (pid_str == NULL)
|
|
goto error;
|
|
|
|
/* XXX: Should it check whether the pid contains embedded
|
|
newlines? */
|
|
if (!PyUnicode_IS_ASCII(pid_str)) {
|
|
PyErr_SetString(_Pickle_GetGlobalState()->PicklingError,
|
|
"persistent IDs in protocol 0 must be "
|
|
"ASCII strings");
|
|
Py_DECREF(pid_str);
|
|
goto error;
|
|
}
|
|
|
|
if (_Pickler_Write(self, &persid_op, 1) < 0 ||
|
|
_Pickler_Write(self, PyUnicode_DATA(pid_str),
|
|
PyUnicode_GET_LENGTH(pid_str)) < 0 ||
|
|
_Pickler_Write(self, "\n", 1) < 0) {
|
|
Py_DECREF(pid_str);
|
|
goto error;
|
|
}
|
|
Py_DECREF(pid_str);
|
|
}
|
|
status = 1;
|
|
}
|
|
|
|
if (0) {
|
|
error:
|
|
status = -1;
|
|
}
|
|
Py_XDECREF(pid);
|
|
|
|
return status;
|
|
}
|
|
|
|
static PyObject *
|
|
get_class(PyObject *obj)
|
|
{
|
|
PyObject *cls;
|
|
_Py_IDENTIFIER(__class__);
|
|
|
|
if (_PyObject_LookupAttrId(obj, &PyId___class__, &cls) == 0) {
|
|
cls = (PyObject *) Py_TYPE(obj);
|
|
Py_INCREF(cls);
|
|
}
|
|
return cls;
|
|
}
|
|
|
|
/* We're saving obj, and args is the 2-thru-5 tuple returned by the
|
|
* appropriate __reduce__ method for obj.
|
|
*/
|
|
static int
|
|
save_reduce(PicklerObject *self, PyObject *args, PyObject *obj)
|
|
{
|
|
PyObject *callable;
|
|
PyObject *argtup;
|
|
PyObject *state = NULL;
|
|
PyObject *listitems = Py_None;
|
|
PyObject *dictitems = Py_None;
|
|
PyObject *state_setter = Py_None;
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
Py_ssize_t size;
|
|
int use_newobj = 0, use_newobj_ex = 0;
|
|
|
|
const char reduce_op = REDUCE;
|
|
const char build_op = BUILD;
|
|
const char newobj_op = NEWOBJ;
|
|
const char newobj_ex_op = NEWOBJ_EX;
|
|
|
|
size = PyTuple_Size(args);
|
|
if (size < 2 || size > 6) {
|
|
PyErr_SetString(st->PicklingError, "tuple returned by "
|
|
"__reduce__ must contain 2 through 6 elements");
|
|
return -1;
|
|
}
|
|
|
|
if (!PyArg_UnpackTuple(args, "save_reduce", 2, 6,
|
|
&callable, &argtup, &state, &listitems, &dictitems,
|
|
&state_setter))
|
|
return -1;
|
|
|
|
if (!PyCallable_Check(callable)) {
|
|
PyErr_SetString(st->PicklingError, "first item of the tuple "
|
|
"returned by __reduce__ must be callable");
|
|
return -1;
|
|
}
|
|
if (!PyTuple_Check(argtup)) {
|
|
PyErr_SetString(st->PicklingError, "second item of the tuple "
|
|
"returned by __reduce__ must be a tuple");
|
|
return -1;
|
|
}
|
|
|
|
if (state == Py_None)
|
|
state = NULL;
|
|
|
|
if (listitems == Py_None)
|
|
listitems = NULL;
|
|
else if (!PyIter_Check(listitems)) {
|
|
PyErr_Format(st->PicklingError, "fourth element of the tuple "
|
|
"returned by __reduce__ must be an iterator, not %s",
|
|
Py_TYPE(listitems)->tp_name);
|
|
return -1;
|
|
}
|
|
|
|
if (dictitems == Py_None)
|
|
dictitems = NULL;
|
|
else if (!PyIter_Check(dictitems)) {
|
|
PyErr_Format(st->PicklingError, "fifth element of the tuple "
|
|
"returned by __reduce__ must be an iterator, not %s",
|
|
Py_TYPE(dictitems)->tp_name);
|
|
return -1;
|
|
}
|
|
|
|
if (state_setter == Py_None)
|
|
state_setter = NULL;
|
|
else if (!PyCallable_Check(state_setter)) {
|
|
PyErr_Format(st->PicklingError, "sixth element of the tuple "
|
|
"returned by __reduce__ must be a function, not %s",
|
|
Py_TYPE(state_setter)->tp_name);
|
|
return -1;
|
|
}
|
|
|
|
if (self->proto >= 2) {
|
|
PyObject *name;
|
|
_Py_IDENTIFIER(__name__);
|
|
|
|
if (_PyObject_LookupAttrId(callable, &PyId___name__, &name) < 0) {
|
|
return -1;
|
|
}
|
|
if (name != NULL && PyUnicode_Check(name)) {
|
|
_Py_IDENTIFIER(__newobj_ex__);
|
|
use_newobj_ex = _PyUnicode_EqualToASCIIId(
|
|
name, &PyId___newobj_ex__);
|
|
if (!use_newobj_ex) {
|
|
_Py_IDENTIFIER(__newobj__);
|
|
use_newobj = _PyUnicode_EqualToASCIIId(name, &PyId___newobj__);
|
|
}
|
|
}
|
|
Py_XDECREF(name);
|
|
}
|
|
|
|
if (use_newobj_ex) {
|
|
PyObject *cls;
|
|
PyObject *args;
|
|
PyObject *kwargs;
|
|
|
|
if (PyTuple_GET_SIZE(argtup) != 3) {
|
|
PyErr_Format(st->PicklingError,
|
|
"length of the NEWOBJ_EX argument tuple must be "
|
|
"exactly 3, not %zd", PyTuple_GET_SIZE(argtup));
|
|
return -1;
|
|
}
|
|
|
|
cls = PyTuple_GET_ITEM(argtup, 0);
|
|
if (!PyType_Check(cls)) {
|
|
PyErr_Format(st->PicklingError,
|
|
"first item from NEWOBJ_EX argument tuple must "
|
|
"be a class, not %.200s", Py_TYPE(cls)->tp_name);
|
|
return -1;
|
|
}
|
|
args = PyTuple_GET_ITEM(argtup, 1);
|
|
if (!PyTuple_Check(args)) {
|
|
PyErr_Format(st->PicklingError,
|
|
"second item from NEWOBJ_EX argument tuple must "
|
|
"be a tuple, not %.200s", Py_TYPE(args)->tp_name);
|
|
return -1;
|
|
}
|
|
kwargs = PyTuple_GET_ITEM(argtup, 2);
|
|
if (!PyDict_Check(kwargs)) {
|
|
PyErr_Format(st->PicklingError,
|
|
"third item from NEWOBJ_EX argument tuple must "
|
|
"be a dict, not %.200s", Py_TYPE(kwargs)->tp_name);
|
|
return -1;
|
|
}
|
|
|
|
if (self->proto >= 4) {
|
|
if (save(self, cls, 0) < 0 ||
|
|
save(self, args, 0) < 0 ||
|
|
save(self, kwargs, 0) < 0 ||
|
|
_Pickler_Write(self, &newobj_ex_op, 1) < 0) {
|
|
return -1;
|
|
}
|
|
}
|
|
else {
|
|
PyObject *newargs;
|
|
PyObject *cls_new;
|
|
Py_ssize_t i;
|
|
_Py_IDENTIFIER(__new__);
|
|
|
|
newargs = PyTuple_New(PyTuple_GET_SIZE(args) + 2);
|
|
if (newargs == NULL)
|
|
return -1;
|
|
|
|
cls_new = _PyObject_GetAttrId(cls, &PyId___new__);
|
|
if (cls_new == NULL) {
|
|
Py_DECREF(newargs);
|
|
return -1;
|
|
}
|
|
PyTuple_SET_ITEM(newargs, 0, cls_new);
|
|
Py_INCREF(cls);
|
|
PyTuple_SET_ITEM(newargs, 1, cls);
|
|
for (i = 0; i < PyTuple_GET_SIZE(args); i++) {
|
|
PyObject *item = PyTuple_GET_ITEM(args, i);
|
|
Py_INCREF(item);
|
|
PyTuple_SET_ITEM(newargs, i + 2, item);
|
|
}
|
|
|
|
callable = PyObject_Call(st->partial, newargs, kwargs);
|
|
Py_DECREF(newargs);
|
|
if (callable == NULL)
|
|
return -1;
|
|
|
|
newargs = PyTuple_New(0);
|
|
if (newargs == NULL) {
|
|
Py_DECREF(callable);
|
|
return -1;
|
|
}
|
|
|
|
if (save(self, callable, 0) < 0 ||
|
|
save(self, newargs, 0) < 0 ||
|
|
_Pickler_Write(self, &reduce_op, 1) < 0) {
|
|
Py_DECREF(newargs);
|
|
Py_DECREF(callable);
|
|
return -1;
|
|
}
|
|
Py_DECREF(newargs);
|
|
Py_DECREF(callable);
|
|
}
|
|
}
|
|
else if (use_newobj) {
|
|
PyObject *cls;
|
|
PyObject *newargtup;
|
|
PyObject *obj_class;
|
|
int p;
|
|
|
|
/* Sanity checks. */
|
|
if (PyTuple_GET_SIZE(argtup) < 1) {
|
|
PyErr_SetString(st->PicklingError, "__newobj__ arglist is empty");
|
|
return -1;
|
|
}
|
|
|
|
cls = PyTuple_GET_ITEM(argtup, 0);
|
|
if (!PyType_Check(cls)) {
|
|
PyErr_SetString(st->PicklingError, "args[0] from "
|
|
"__newobj__ args is not a type");
|
|
return -1;
|
|
}
|
|
|
|
if (obj != NULL) {
|
|
obj_class = get_class(obj);
|
|
if (obj_class == NULL) {
|
|
return -1;
|
|
}
|
|
p = obj_class != cls;
|
|
Py_DECREF(obj_class);
|
|
if (p) {
|
|
PyErr_SetString(st->PicklingError, "args[0] from "
|
|
"__newobj__ args has the wrong class");
|
|
return -1;
|
|
}
|
|
}
|
|
/* XXX: These calls save() are prone to infinite recursion. Imagine
|
|
what happen if the value returned by the __reduce__() method of
|
|
some extension type contains another object of the same type. Ouch!
|
|
|
|
Here is a quick example, that I ran into, to illustrate what I
|
|
mean:
|
|
|
|
>>> import pickle, copyreg
|
|
>>> copyreg.dispatch_table.pop(complex)
|
|
>>> pickle.dumps(1+2j)
|
|
Traceback (most recent call last):
|
|
...
|
|
RecursionError: maximum recursion depth exceeded
|
|
|
|
Removing the complex class from copyreg.dispatch_table made the
|
|
__reduce_ex__() method emit another complex object:
|
|
|
|
>>> (1+1j).__reduce_ex__(2)
|
|
(<function __newobj__ at 0xb7b71c3c>,
|
|
(<class 'complex'>, (1+1j)), None, None, None)
|
|
|
|
Thus when save() was called on newargstup (the 2nd item) recursion
|
|
ensued. Of course, the bug was in the complex class which had a
|
|
broken __getnewargs__() that emitted another complex object. But,
|
|
the point, here, is it is quite easy to end up with a broken reduce
|
|
function. */
|
|
|
|
/* Save the class and its __new__ arguments. */
|
|
if (save(self, cls, 0) < 0)
|
|
return -1;
|
|
|
|
newargtup = PyTuple_GetSlice(argtup, 1, PyTuple_GET_SIZE(argtup));
|
|
if (newargtup == NULL)
|
|
return -1;
|
|
|
|
p = save(self, newargtup, 0);
|
|
Py_DECREF(newargtup);
|
|
if (p < 0)
|
|
return -1;
|
|
|
|
/* Add NEWOBJ opcode. */
|
|
if (_Pickler_Write(self, &newobj_op, 1) < 0)
|
|
return -1;
|
|
}
|
|
else { /* Not using NEWOBJ. */
|
|
if (save(self, callable, 0) < 0 ||
|
|
save(self, argtup, 0) < 0 ||
|
|
_Pickler_Write(self, &reduce_op, 1) < 0)
|
|
return -1;
|
|
}
|
|
|
|
/* obj can be NULL when save_reduce() is used directly. A NULL obj means
|
|
the caller do not want to memoize the object. Not particularly useful,
|
|
but that is to mimic the behavior save_reduce() in pickle.py when
|
|
obj is None. */
|
|
if (obj != NULL) {
|
|
/* If the object is already in the memo, this means it is
|
|
recursive. In this case, throw away everything we put on the
|
|
stack, and fetch the object back from the memo. */
|
|
if (PyMemoTable_Get(self->memo, obj)) {
|
|
const char pop_op = POP;
|
|
|
|
if (_Pickler_Write(self, &pop_op, 1) < 0)
|
|
return -1;
|
|
if (memo_get(self, obj) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
else if (memo_put(self, obj) < 0)
|
|
return -1;
|
|
}
|
|
|
|
if (listitems && batch_list(self, listitems) < 0)
|
|
return -1;
|
|
|
|
if (dictitems && batch_dict(self, dictitems) < 0)
|
|
return -1;
|
|
|
|
if (state) {
|
|
if (state_setter == NULL) {
|
|
if (save(self, state, 0) < 0 ||
|
|
_Pickler_Write(self, &build_op, 1) < 0)
|
|
return -1;
|
|
}
|
|
else {
|
|
|
|
/* If a state_setter is specified, call it instead of load_build to
|
|
* update obj's with its previous state.
|
|
* The first 4 save/write instructions push state_setter and its
|
|
* tuple of expected arguments (obj, state) onto the stack. The
|
|
* REDUCE opcode triggers the state_setter(obj, state) function
|
|
* call. Finally, because state-updating routines only do in-place
|
|
* modification, the whole operation has to be stack-transparent.
|
|
* Thus, we finally pop the call's output from the stack.*/
|
|
|
|
const char tupletwo_op = TUPLE2;
|
|
const char pop_op = POP;
|
|
if (save(self, state_setter, 0) < 0 ||
|
|
save(self, obj, 0) < 0 || save(self, state, 0) < 0 ||
|
|
_Pickler_Write(self, &tupletwo_op, 1) < 0 ||
|
|
_Pickler_Write(self, &reduce_op, 1) < 0 ||
|
|
_Pickler_Write(self, &pop_op, 1) < 0)
|
|
return -1;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
save(PicklerObject *self, PyObject *obj, int pers_save)
|
|
{
|
|
PyTypeObject *type;
|
|
PyObject *reduce_func = NULL;
|
|
PyObject *reduce_value = NULL;
|
|
int status = 0;
|
|
|
|
if (_Pickler_OpcodeBoundary(self) < 0)
|
|
return -1;
|
|
|
|
/* The extra pers_save argument is necessary to avoid calling save_pers()
|
|
on its returned object. */
|
|
if (!pers_save && self->pers_func) {
|
|
/* save_pers() returns:
|
|
-1 to signal an error;
|
|
0 if it did nothing successfully;
|
|
1 if a persistent id was saved.
|
|
*/
|
|
if ((status = save_pers(self, obj)) != 0)
|
|
return status;
|
|
}
|
|
|
|
type = Py_TYPE(obj);
|
|
|
|
/* The old cPickle had an optimization that used switch-case statement
|
|
dispatching on the first letter of the type name. This has was removed
|
|
since benchmarks shown that this optimization was actually slowing
|
|
things down. */
|
|
|
|
/* Atom types; these aren't memoized, so don't check the memo. */
|
|
|
|
if (obj == Py_None) {
|
|
return save_none(self, obj);
|
|
}
|
|
else if (obj == Py_False || obj == Py_True) {
|
|
return save_bool(self, obj);
|
|
}
|
|
else if (type == &PyLong_Type) {
|
|
return save_long(self, obj);
|
|
}
|
|
else if (type == &PyFloat_Type) {
|
|
return save_float(self, obj);
|
|
}
|
|
|
|
/* Check the memo to see if it has the object. If so, generate
|
|
a GET (or BINGET) opcode, instead of pickling the object
|
|
once again. */
|
|
if (PyMemoTable_Get(self->memo, obj)) {
|
|
return memo_get(self, obj);
|
|
}
|
|
|
|
if (type == &PyBytes_Type) {
|
|
return save_bytes(self, obj);
|
|
}
|
|
else if (type == &PyUnicode_Type) {
|
|
return save_unicode(self, obj);
|
|
}
|
|
|
|
/* We're only calling Py_EnterRecursiveCall here so that atomic
|
|
types above are pickled faster. */
|
|
if (Py_EnterRecursiveCall(" while pickling an object")) {
|
|
return -1;
|
|
}
|
|
|
|
if (type == &PyDict_Type) {
|
|
status = save_dict(self, obj);
|
|
goto done;
|
|
}
|
|
else if (type == &PySet_Type) {
|
|
status = save_set(self, obj);
|
|
goto done;
|
|
}
|
|
else if (type == &PyFrozenSet_Type) {
|
|
status = save_frozenset(self, obj);
|
|
goto done;
|
|
}
|
|
else if (type == &PyList_Type) {
|
|
status = save_list(self, obj);
|
|
goto done;
|
|
}
|
|
else if (type == &PyTuple_Type) {
|
|
status = save_tuple(self, obj);
|
|
goto done;
|
|
}
|
|
else if (type == &PyByteArray_Type) {
|
|
status = save_bytearray(self, obj);
|
|
goto done;
|
|
}
|
|
else if (type == &PyPickleBuffer_Type) {
|
|
status = save_picklebuffer(self, obj);
|
|
goto done;
|
|
}
|
|
|
|
/* Now, check reducer_override. If it returns NotImplemented,
|
|
* fallback to save_type or save_global, and then perhaps to the
|
|
* regular reduction mechanism.
|
|
*/
|
|
if (self->reducer_override != NULL) {
|
|
reduce_value = PyObject_CallOneArg(self->reducer_override, obj);
|
|
if (reduce_value == NULL) {
|
|
goto error;
|
|
}
|
|
if (reduce_value != Py_NotImplemented) {
|
|
goto reduce;
|
|
}
|
|
Py_DECREF(reduce_value);
|
|
reduce_value = NULL;
|
|
}
|
|
|
|
if (type == &PyType_Type) {
|
|
status = save_type(self, obj);
|
|
goto done;
|
|
}
|
|
else if (type == &PyFunction_Type) {
|
|
status = save_global(self, obj, NULL);
|
|
goto done;
|
|
}
|
|
|
|
/* XXX: This part needs some unit tests. */
|
|
|
|
/* Get a reduction callable, and call it. This may come from
|
|
* self.dispatch_table, copyreg.dispatch_table, the object's
|
|
* __reduce_ex__ method, or the object's __reduce__ method.
|
|
*/
|
|
if (self->dispatch_table == NULL) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
reduce_func = PyDict_GetItemWithError(st->dispatch_table,
|
|
(PyObject *)type);
|
|
if (reduce_func == NULL) {
|
|
if (PyErr_Occurred()) {
|
|
goto error;
|
|
}
|
|
} else {
|
|
/* PyDict_GetItemWithError() returns a borrowed reference.
|
|
Increase the reference count to be consistent with
|
|
PyObject_GetItem and _PyObject_GetAttrId used below. */
|
|
Py_INCREF(reduce_func);
|
|
}
|
|
} else {
|
|
reduce_func = PyObject_GetItem(self->dispatch_table,
|
|
(PyObject *)type);
|
|
if (reduce_func == NULL) {
|
|
if (PyErr_ExceptionMatches(PyExc_KeyError))
|
|
PyErr_Clear();
|
|
else
|
|
goto error;
|
|
}
|
|
}
|
|
if (reduce_func != NULL) {
|
|
Py_INCREF(obj);
|
|
reduce_value = _Pickle_FastCall(reduce_func, obj);
|
|
}
|
|
else if (PyType_IsSubtype(type, &PyType_Type)) {
|
|
status = save_global(self, obj, NULL);
|
|
goto done;
|
|
}
|
|
else {
|
|
_Py_IDENTIFIER(__reduce__);
|
|
_Py_IDENTIFIER(__reduce_ex__);
|
|
|
|
/* XXX: If the __reduce__ method is defined, __reduce_ex__ is
|
|
automatically defined as __reduce__. While this is convenient, this
|
|
make it impossible to know which method was actually called. Of
|
|
course, this is not a big deal. But still, it would be nice to let
|
|
the user know which method was called when something go
|
|
wrong. Incidentally, this means if __reduce_ex__ is not defined, we
|
|
don't actually have to check for a __reduce__ method. */
|
|
|
|
/* Check for a __reduce_ex__ method. */
|
|
if (_PyObject_LookupAttrId(obj, &PyId___reduce_ex__, &reduce_func) < 0) {
|
|
goto error;
|
|
}
|
|
if (reduce_func != NULL) {
|
|
PyObject *proto;
|
|
proto = PyLong_FromLong(self->proto);
|
|
if (proto != NULL) {
|
|
reduce_value = _Pickle_FastCall(reduce_func, proto);
|
|
}
|
|
}
|
|
else {
|
|
/* Check for a __reduce__ method. */
|
|
if (_PyObject_LookupAttrId(obj, &PyId___reduce__, &reduce_func) < 0) {
|
|
goto error;
|
|
}
|
|
if (reduce_func != NULL) {
|
|
reduce_value = PyObject_CallNoArgs(reduce_func);
|
|
}
|
|
else {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_Format(st->PicklingError,
|
|
"can't pickle '%.200s' object: %R",
|
|
type->tp_name, obj);
|
|
goto error;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (reduce_value == NULL)
|
|
goto error;
|
|
|
|
reduce:
|
|
if (PyUnicode_Check(reduce_value)) {
|
|
status = save_global(self, obj, reduce_value);
|
|
goto done;
|
|
}
|
|
|
|
if (!PyTuple_Check(reduce_value)) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->PicklingError,
|
|
"__reduce__ must return a string or tuple");
|
|
goto error;
|
|
}
|
|
|
|
status = save_reduce(self, reduce_value, obj);
|
|
|
|
if (0) {
|
|
error:
|
|
status = -1;
|
|
}
|
|
done:
|
|
|
|
Py_LeaveRecursiveCall();
|
|
Py_XDECREF(reduce_func);
|
|
Py_XDECREF(reduce_value);
|
|
|
|
return status;
|
|
}
|
|
|
|
static int
|
|
dump(PicklerObject *self, PyObject *obj)
|
|
{
|
|
const char stop_op = STOP;
|
|
int status = -1;
|
|
PyObject *tmp;
|
|
_Py_IDENTIFIER(reducer_override);
|
|
|
|
if (_PyObject_LookupAttrId((PyObject *)self, &PyId_reducer_override,
|
|
&tmp) < 0) {
|
|
goto error;
|
|
}
|
|
/* Cache the reducer_override method, if it exists. */
|
|
if (tmp != NULL) {
|
|
Py_XSETREF(self->reducer_override, tmp);
|
|
}
|
|
else {
|
|
Py_CLEAR(self->reducer_override);
|
|
}
|
|
|
|
if (self->proto >= 2) {
|
|
char header[2];
|
|
|
|
header[0] = PROTO;
|
|
assert(self->proto >= 0 && self->proto < 256);
|
|
header[1] = (unsigned char)self->proto;
|
|
if (_Pickler_Write(self, header, 2) < 0)
|
|
goto error;
|
|
if (self->proto >= 4)
|
|
self->framing = 1;
|
|
}
|
|
|
|
if (save(self, obj, 0) < 0 ||
|
|
_Pickler_Write(self, &stop_op, 1) < 0 ||
|
|
_Pickler_CommitFrame(self) < 0)
|
|
goto error;
|
|
|
|
// Success
|
|
status = 0;
|
|
|
|
error:
|
|
self->framing = 0;
|
|
|
|
/* Break the reference cycle we generated at the beginning this function
|
|
* call when setting the reducer_override attribute of the Pickler instance
|
|
* to a bound method of the same instance. This is important as the Pickler
|
|
* instance holds a reference to each object it has pickled (through its
|
|
* memo): thus, these objects wont be garbage-collected as long as the
|
|
* Pickler itself is not collected. */
|
|
Py_CLEAR(self->reducer_override);
|
|
return status;
|
|
}
|
|
|
|
/*[clinic input]
|
|
|
|
_pickle.Pickler.clear_memo
|
|
|
|
Clears the pickler's "memo".
|
|
|
|
The memo is the data structure that remembers which objects the
|
|
pickler has already seen, so that shared or recursive objects are
|
|
pickled by reference and not by value. This method is useful when
|
|
re-using picklers.
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_Pickler_clear_memo_impl(PicklerObject *self)
|
|
/*[clinic end generated code: output=8665c8658aaa094b input=01bdad52f3d93e56]*/
|
|
{
|
|
if (self->memo)
|
|
PyMemoTable_Clear(self->memo);
|
|
|
|
Py_RETURN_NONE;
|
|
}
|
|
|
|
/*[clinic input]
|
|
|
|
_pickle.Pickler.dump
|
|
|
|
obj: object
|
|
/
|
|
|
|
Write a pickled representation of the given object to the open file.
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_Pickler_dump(PicklerObject *self, PyObject *obj)
|
|
/*[clinic end generated code: output=87ecad1261e02ac7 input=552eb1c0f52260d9]*/
|
|
{
|
|
/* Check whether the Pickler was initialized correctly (issue3664).
|
|
Developers often forget to call __init__() in their subclasses, which
|
|
would trigger a segfault without this check. */
|
|
if (self->write == NULL) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_Format(st->PicklingError,
|
|
"Pickler.__init__() was not called by %s.__init__()",
|
|
Py_TYPE(self)->tp_name);
|
|
return NULL;
|
|
}
|
|
|
|
if (_Pickler_ClearBuffer(self) < 0)
|
|
return NULL;
|
|
|
|
if (dump(self, obj) < 0)
|
|
return NULL;
|
|
|
|
if (_Pickler_FlushToFile(self) < 0)
|
|
return NULL;
|
|
|
|
Py_RETURN_NONE;
|
|
}
|
|
|
|
/*[clinic input]
|
|
|
|
_pickle.Pickler.__sizeof__ -> Py_ssize_t
|
|
|
|
Returns size in memory, in bytes.
|
|
[clinic start generated code]*/
|
|
|
|
static Py_ssize_t
|
|
_pickle_Pickler___sizeof___impl(PicklerObject *self)
|
|
/*[clinic end generated code: output=106edb3123f332e1 input=8cbbec9bd5540d42]*/
|
|
{
|
|
Py_ssize_t res, s;
|
|
|
|
res = _PyObject_SIZE(Py_TYPE(self));
|
|
if (self->memo != NULL) {
|
|
res += sizeof(PyMemoTable);
|
|
res += self->memo->mt_allocated * sizeof(PyMemoEntry);
|
|
}
|
|
if (self->output_buffer != NULL) {
|
|
s = _PySys_GetSizeOf(self->output_buffer);
|
|
if (s == -1)
|
|
return -1;
|
|
res += s;
|
|
}
|
|
return res;
|
|
}
|
|
|
|
static struct PyMethodDef Pickler_methods[] = {
|
|
_PICKLE_PICKLER_DUMP_METHODDEF
|
|
_PICKLE_PICKLER_CLEAR_MEMO_METHODDEF
|
|
_PICKLE_PICKLER___SIZEOF___METHODDEF
|
|
{NULL, NULL} /* sentinel */
|
|
};
|
|
|
|
static void
|
|
Pickler_dealloc(PicklerObject *self)
|
|
{
|
|
PyObject_GC_UnTrack(self);
|
|
|
|
Py_XDECREF(self->output_buffer);
|
|
Py_XDECREF(self->write);
|
|
Py_XDECREF(self->pers_func);
|
|
Py_XDECREF(self->dispatch_table);
|
|
Py_XDECREF(self->fast_memo);
|
|
Py_XDECREF(self->reducer_override);
|
|
Py_XDECREF(self->buffer_callback);
|
|
|
|
PyMemoTable_Del(self->memo);
|
|
|
|
Py_TYPE(self)->tp_free((PyObject *)self);
|
|
}
|
|
|
|
static int
|
|
Pickler_traverse(PicklerObject *self, visitproc visit, void *arg)
|
|
{
|
|
Py_VISIT(self->write);
|
|
Py_VISIT(self->pers_func);
|
|
Py_VISIT(self->dispatch_table);
|
|
Py_VISIT(self->fast_memo);
|
|
Py_VISIT(self->reducer_override);
|
|
Py_VISIT(self->buffer_callback);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
Pickler_clear(PicklerObject *self)
|
|
{
|
|
Py_CLEAR(self->output_buffer);
|
|
Py_CLEAR(self->write);
|
|
Py_CLEAR(self->pers_func);
|
|
Py_CLEAR(self->dispatch_table);
|
|
Py_CLEAR(self->fast_memo);
|
|
Py_CLEAR(self->reducer_override);
|
|
Py_CLEAR(self->buffer_callback);
|
|
|
|
if (self->memo != NULL) {
|
|
PyMemoTable *memo = self->memo;
|
|
self->memo = NULL;
|
|
PyMemoTable_Del(memo);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
|
|
/*[clinic input]
|
|
|
|
_pickle.Pickler.__init__
|
|
|
|
file: object
|
|
protocol: object = None
|
|
fix_imports: bool = True
|
|
buffer_callback: object = None
|
|
|
|
This takes a binary file for writing a pickle data stream.
|
|
|
|
The optional *protocol* argument tells the pickler to use the given
|
|
protocol; supported protocols are 0, 1, 2, 3, 4 and 5. The default
|
|
protocol is 4. It was introduced in Python 3.4, and is incompatible
|
|
with previous versions.
|
|
|
|
Specifying a negative protocol version selects the highest protocol
|
|
version supported. The higher the protocol used, the more recent the
|
|
version of Python needed to read the pickle produced.
|
|
|
|
The *file* argument must have a write() method that accepts a single
|
|
bytes argument. It can thus be a file object opened for binary
|
|
writing, an io.BytesIO instance, or any other custom object that meets
|
|
this interface.
|
|
|
|
If *fix_imports* is True and protocol is less than 3, pickle will try
|
|
to map the new Python 3 names to the old module names used in Python
|
|
2, so that the pickle data stream is readable with Python 2.
|
|
|
|
If *buffer_callback* is None (the default), buffer views are
|
|
serialized into *file* as part of the pickle stream.
|
|
|
|
If *buffer_callback* is not None, then it can be called any number
|
|
of times with a buffer view. If the callback returns a false value
|
|
(such as None), the given buffer is out-of-band; otherwise the
|
|
buffer is serialized in-band, i.e. inside the pickle stream.
|
|
|
|
It is an error if *buffer_callback* is not None and *protocol*
|
|
is None or smaller than 5.
|
|
|
|
[clinic start generated code]*/
|
|
|
|
static int
|
|
_pickle_Pickler___init___impl(PicklerObject *self, PyObject *file,
|
|
PyObject *protocol, int fix_imports,
|
|
PyObject *buffer_callback)
|
|
/*[clinic end generated code: output=0abedc50590d259b input=a7c969699bf5dad3]*/
|
|
{
|
|
_Py_IDENTIFIER(persistent_id);
|
|
_Py_IDENTIFIER(dispatch_table);
|
|
|
|
/* In case of multiple __init__() calls, clear previous content. */
|
|
if (self->write != NULL)
|
|
(void)Pickler_clear(self);
|
|
|
|
if (_Pickler_SetProtocol(self, protocol, fix_imports) < 0)
|
|
return -1;
|
|
|
|
if (_Pickler_SetOutputStream(self, file) < 0)
|
|
return -1;
|
|
|
|
if (_Pickler_SetBufferCallback(self, buffer_callback) < 0)
|
|
return -1;
|
|
|
|
/* memo and output_buffer may have already been created in _Pickler_New */
|
|
if (self->memo == NULL) {
|
|
self->memo = PyMemoTable_New();
|
|
if (self->memo == NULL)
|
|
return -1;
|
|
}
|
|
self->output_len = 0;
|
|
if (self->output_buffer == NULL) {
|
|
self->max_output_len = WRITE_BUF_SIZE;
|
|
self->output_buffer = PyBytes_FromStringAndSize(NULL,
|
|
self->max_output_len);
|
|
if (self->output_buffer == NULL)
|
|
return -1;
|
|
}
|
|
|
|
self->fast = 0;
|
|
self->fast_nesting = 0;
|
|
self->fast_memo = NULL;
|
|
|
|
if (init_method_ref((PyObject *)self, &PyId_persistent_id,
|
|
&self->pers_func, &self->pers_func_self) < 0)
|
|
{
|
|
return -1;
|
|
}
|
|
|
|
if (_PyObject_LookupAttrId((PyObject *)self,
|
|
&PyId_dispatch_table, &self->dispatch_table) < 0) {
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
/* Define a proxy object for the Pickler's internal memo object. This is to
|
|
* avoid breaking code like:
|
|
* pickler.memo.clear()
|
|
* and
|
|
* pickler.memo = saved_memo
|
|
* Is this a good idea? Not really, but we don't want to break code that uses
|
|
* it. Note that we don't implement the entire mapping API here. This is
|
|
* intentional, as these should be treated as black-box implementation details.
|
|
*/
|
|
|
|
/*[clinic input]
|
|
_pickle.PicklerMemoProxy.clear
|
|
|
|
Remove all items from memo.
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_PicklerMemoProxy_clear_impl(PicklerMemoProxyObject *self)
|
|
/*[clinic end generated code: output=5fb9370d48ae8b05 input=ccc186dacd0f1405]*/
|
|
{
|
|
if (self->pickler->memo)
|
|
PyMemoTable_Clear(self->pickler->memo);
|
|
Py_RETURN_NONE;
|
|
}
|
|
|
|
/*[clinic input]
|
|
_pickle.PicklerMemoProxy.copy
|
|
|
|
Copy the memo to a new object.
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_PicklerMemoProxy_copy_impl(PicklerMemoProxyObject *self)
|
|
/*[clinic end generated code: output=bb83a919d29225ef input=b73043485ac30b36]*/
|
|
{
|
|
PyMemoTable *memo;
|
|
PyObject *new_memo = PyDict_New();
|
|
if (new_memo == NULL)
|
|
return NULL;
|
|
|
|
memo = self->pickler->memo;
|
|
for (size_t i = 0; i < memo->mt_allocated; ++i) {
|
|
PyMemoEntry entry = memo->mt_table[i];
|
|
if (entry.me_key != NULL) {
|
|
int status;
|
|
PyObject *key, *value;
|
|
|
|
key = PyLong_FromVoidPtr(entry.me_key);
|
|
value = Py_BuildValue("nO", entry.me_value, entry.me_key);
|
|
|
|
if (key == NULL || value == NULL) {
|
|
Py_XDECREF(key);
|
|
Py_XDECREF(value);
|
|
goto error;
|
|
}
|
|
status = PyDict_SetItem(new_memo, key, value);
|
|
Py_DECREF(key);
|
|
Py_DECREF(value);
|
|
if (status < 0)
|
|
goto error;
|
|
}
|
|
}
|
|
return new_memo;
|
|
|
|
error:
|
|
Py_XDECREF(new_memo);
|
|
return NULL;
|
|
}
|
|
|
|
/*[clinic input]
|
|
_pickle.PicklerMemoProxy.__reduce__
|
|
|
|
Implement pickle support.
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_PicklerMemoProxy___reduce___impl(PicklerMemoProxyObject *self)
|
|
/*[clinic end generated code: output=bebba1168863ab1d input=2f7c540e24b7aae4]*/
|
|
{
|
|
PyObject *reduce_value, *dict_args;
|
|
PyObject *contents = _pickle_PicklerMemoProxy_copy_impl(self);
|
|
if (contents == NULL)
|
|
return NULL;
|
|
|
|
reduce_value = PyTuple_New(2);
|
|
if (reduce_value == NULL) {
|
|
Py_DECREF(contents);
|
|
return NULL;
|
|
}
|
|
dict_args = PyTuple_New(1);
|
|
if (dict_args == NULL) {
|
|
Py_DECREF(contents);
|
|
Py_DECREF(reduce_value);
|
|
return NULL;
|
|
}
|
|
PyTuple_SET_ITEM(dict_args, 0, contents);
|
|
Py_INCREF((PyObject *)&PyDict_Type);
|
|
PyTuple_SET_ITEM(reduce_value, 0, (PyObject *)&PyDict_Type);
|
|
PyTuple_SET_ITEM(reduce_value, 1, dict_args);
|
|
return reduce_value;
|
|
}
|
|
|
|
static PyMethodDef picklerproxy_methods[] = {
|
|
_PICKLE_PICKLERMEMOPROXY_CLEAR_METHODDEF
|
|
_PICKLE_PICKLERMEMOPROXY_COPY_METHODDEF
|
|
_PICKLE_PICKLERMEMOPROXY___REDUCE___METHODDEF
|
|
{NULL, NULL} /* sentinel */
|
|
};
|
|
|
|
static void
|
|
PicklerMemoProxy_dealloc(PicklerMemoProxyObject *self)
|
|
{
|
|
PyObject_GC_UnTrack(self);
|
|
Py_XDECREF(self->pickler);
|
|
PyObject_GC_Del((PyObject *)self);
|
|
}
|
|
|
|
static int
|
|
PicklerMemoProxy_traverse(PicklerMemoProxyObject *self,
|
|
visitproc visit, void *arg)
|
|
{
|
|
Py_VISIT(self->pickler);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
PicklerMemoProxy_clear(PicklerMemoProxyObject *self)
|
|
{
|
|
Py_CLEAR(self->pickler);
|
|
return 0;
|
|
}
|
|
|
|
static PyTypeObject PicklerMemoProxyType = {
|
|
PyVarObject_HEAD_INIT(NULL, 0)
|
|
"_pickle.PicklerMemoProxy", /*tp_name*/
|
|
sizeof(PicklerMemoProxyObject), /*tp_basicsize*/
|
|
0,
|
|
(destructor)PicklerMemoProxy_dealloc, /* tp_dealloc */
|
|
0, /* tp_vectorcall_offset */
|
|
0, /* tp_getattr */
|
|
0, /* tp_setattr */
|
|
0, /* tp_as_async */
|
|
0, /* tp_repr */
|
|
0, /* tp_as_number */
|
|
0, /* tp_as_sequence */
|
|
0, /* tp_as_mapping */
|
|
PyObject_HashNotImplemented, /* tp_hash */
|
|
0, /* tp_call */
|
|
0, /* tp_str */
|
|
PyObject_GenericGetAttr, /* tp_getattro */
|
|
PyObject_GenericSetAttr, /* tp_setattro */
|
|
0, /* tp_as_buffer */
|
|
Py_TPFLAGS_DEFAULT | Py_TPFLAGS_BASETYPE | Py_TPFLAGS_HAVE_GC,
|
|
0, /* tp_doc */
|
|
(traverseproc)PicklerMemoProxy_traverse, /* tp_traverse */
|
|
(inquiry)PicklerMemoProxy_clear, /* tp_clear */
|
|
0, /* tp_richcompare */
|
|
0, /* tp_weaklistoffset */
|
|
0, /* tp_iter */
|
|
0, /* tp_iternext */
|
|
picklerproxy_methods, /* tp_methods */
|
|
};
|
|
|
|
static PyObject *
|
|
PicklerMemoProxy_New(PicklerObject *pickler)
|
|
{
|
|
PicklerMemoProxyObject *self;
|
|
|
|
self = PyObject_GC_New(PicklerMemoProxyObject, &PicklerMemoProxyType);
|
|
if (self == NULL)
|
|
return NULL;
|
|
Py_INCREF(pickler);
|
|
self->pickler = pickler;
|
|
PyObject_GC_Track(self);
|
|
return (PyObject *)self;
|
|
}
|
|
|
|
/*****************************************************************************/
|
|
|
|
static PyObject *
|
|
Pickler_get_memo(PicklerObject *self, void *Py_UNUSED(ignored))
|
|
{
|
|
return PicklerMemoProxy_New(self);
|
|
}
|
|
|
|
static int
|
|
Pickler_set_memo(PicklerObject *self, PyObject *obj, void *Py_UNUSED(ignored))
|
|
{
|
|
PyMemoTable *new_memo = NULL;
|
|
|
|
if (obj == NULL) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"attribute deletion is not supported");
|
|
return -1;
|
|
}
|
|
|
|
if (Py_IS_TYPE(obj, &PicklerMemoProxyType)) {
|
|
PicklerObject *pickler =
|
|
((PicklerMemoProxyObject *)obj)->pickler;
|
|
|
|
new_memo = PyMemoTable_Copy(pickler->memo);
|
|
if (new_memo == NULL)
|
|
return -1;
|
|
}
|
|
else if (PyDict_Check(obj)) {
|
|
Py_ssize_t i = 0;
|
|
PyObject *key, *value;
|
|
|
|
new_memo = PyMemoTable_New();
|
|
if (new_memo == NULL)
|
|
return -1;
|
|
|
|
while (PyDict_Next(obj, &i, &key, &value)) {
|
|
Py_ssize_t memo_id;
|
|
PyObject *memo_obj;
|
|
|
|
if (!PyTuple_Check(value) || PyTuple_GET_SIZE(value) != 2) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"'memo' values must be 2-item tuples");
|
|
goto error;
|
|
}
|
|
memo_id = PyLong_AsSsize_t(PyTuple_GET_ITEM(value, 0));
|
|
if (memo_id == -1 && PyErr_Occurred())
|
|
goto error;
|
|
memo_obj = PyTuple_GET_ITEM(value, 1);
|
|
if (PyMemoTable_Set(new_memo, memo_obj, memo_id) < 0)
|
|
goto error;
|
|
}
|
|
}
|
|
else {
|
|
PyErr_Format(PyExc_TypeError,
|
|
"'memo' attribute must be a PicklerMemoProxy object "
|
|
"or dict, not %.200s", Py_TYPE(obj)->tp_name);
|
|
return -1;
|
|
}
|
|
|
|
PyMemoTable_Del(self->memo);
|
|
self->memo = new_memo;
|
|
|
|
return 0;
|
|
|
|
error:
|
|
if (new_memo)
|
|
PyMemoTable_Del(new_memo);
|
|
return -1;
|
|
}
|
|
|
|
static PyObject *
|
|
Pickler_get_persid(PicklerObject *self, void *Py_UNUSED(ignored))
|
|
{
|
|
if (self->pers_func == NULL) {
|
|
PyErr_SetString(PyExc_AttributeError, "persistent_id");
|
|
return NULL;
|
|
}
|
|
return reconstruct_method(self->pers_func, self->pers_func_self);
|
|
}
|
|
|
|
static int
|
|
Pickler_set_persid(PicklerObject *self, PyObject *value, void *Py_UNUSED(ignored))
|
|
{
|
|
if (value == NULL) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"attribute deletion is not supported");
|
|
return -1;
|
|
}
|
|
if (!PyCallable_Check(value)) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"persistent_id must be a callable taking one argument");
|
|
return -1;
|
|
}
|
|
|
|
self->pers_func_self = NULL;
|
|
Py_INCREF(value);
|
|
Py_XSETREF(self->pers_func, value);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static PyMemberDef Pickler_members[] = {
|
|
{"bin", T_INT, offsetof(PicklerObject, bin)},
|
|
{"fast", T_INT, offsetof(PicklerObject, fast)},
|
|
{"dispatch_table", T_OBJECT_EX, offsetof(PicklerObject, dispatch_table)},
|
|
{NULL}
|
|
};
|
|
|
|
static PyGetSetDef Pickler_getsets[] = {
|
|
{"memo", (getter)Pickler_get_memo,
|
|
(setter)Pickler_set_memo},
|
|
{"persistent_id", (getter)Pickler_get_persid,
|
|
(setter)Pickler_set_persid},
|
|
{NULL}
|
|
};
|
|
|
|
static PyTypeObject Pickler_Type = {
|
|
PyVarObject_HEAD_INIT(NULL, 0)
|
|
"_pickle.Pickler" , /*tp_name*/
|
|
sizeof(PicklerObject), /*tp_basicsize*/
|
|
0, /*tp_itemsize*/
|
|
(destructor)Pickler_dealloc, /*tp_dealloc*/
|
|
0, /*tp_vectorcall_offset*/
|
|
0, /*tp_getattr*/
|
|
0, /*tp_setattr*/
|
|
0, /*tp_as_async*/
|
|
0, /*tp_repr*/
|
|
0, /*tp_as_number*/
|
|
0, /*tp_as_sequence*/
|
|
0, /*tp_as_mapping*/
|
|
0, /*tp_hash*/
|
|
0, /*tp_call*/
|
|
0, /*tp_str*/
|
|
0, /*tp_getattro*/
|
|
0, /*tp_setattro*/
|
|
0, /*tp_as_buffer*/
|
|
Py_TPFLAGS_DEFAULT | Py_TPFLAGS_BASETYPE | Py_TPFLAGS_HAVE_GC,
|
|
_pickle_Pickler___init____doc__, /*tp_doc*/
|
|
(traverseproc)Pickler_traverse, /*tp_traverse*/
|
|
(inquiry)Pickler_clear, /*tp_clear*/
|
|
0, /*tp_richcompare*/
|
|
0, /*tp_weaklistoffset*/
|
|
0, /*tp_iter*/
|
|
0, /*tp_iternext*/
|
|
Pickler_methods, /*tp_methods*/
|
|
Pickler_members, /*tp_members*/
|
|
Pickler_getsets, /*tp_getset*/
|
|
0, /*tp_base*/
|
|
0, /*tp_dict*/
|
|
0, /*tp_descr_get*/
|
|
0, /*tp_descr_set*/
|
|
0, /*tp_dictoffset*/
|
|
_pickle_Pickler___init__, /*tp_init*/
|
|
PyType_GenericAlloc, /*tp_alloc*/
|
|
PyType_GenericNew, /*tp_new*/
|
|
PyObject_GC_Del, /*tp_free*/
|
|
0, /*tp_is_gc*/
|
|
};
|
|
|
|
/* Temporary helper for calling self.find_class().
|
|
|
|
XXX: It would be nice to able to avoid Python function call overhead, by
|
|
using directly the C version of find_class(), when find_class() is not
|
|
overridden by a subclass. Although, this could become rather hackish. A
|
|
simpler optimization would be to call the C function when self is not a
|
|
subclass instance. */
|
|
static PyObject *
|
|
find_class(UnpicklerObject *self, PyObject *module_name, PyObject *global_name)
|
|
{
|
|
_Py_IDENTIFIER(find_class);
|
|
|
|
return _PyObject_CallMethodIdObjArgs((PyObject *)self, &PyId_find_class,
|
|
module_name, global_name, NULL);
|
|
}
|
|
|
|
static Py_ssize_t
|
|
marker(UnpicklerObject *self)
|
|
{
|
|
Py_ssize_t mark;
|
|
|
|
if (self->num_marks < 1) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->UnpicklingError, "could not find MARK");
|
|
return -1;
|
|
}
|
|
|
|
mark = self->marks[--self->num_marks];
|
|
self->stack->mark_set = self->num_marks != 0;
|
|
self->stack->fence = self->num_marks ?
|
|
self->marks[self->num_marks - 1] : 0;
|
|
return mark;
|
|
}
|
|
|
|
static int
|
|
load_none(UnpicklerObject *self)
|
|
{
|
|
PDATA_APPEND(self->stack, Py_None, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_int(UnpicklerObject *self)
|
|
{
|
|
PyObject *value;
|
|
char *endptr, *s;
|
|
Py_ssize_t len;
|
|
long x;
|
|
|
|
if ((len = _Unpickler_Readline(self, &s)) < 0)
|
|
return -1;
|
|
if (len < 2)
|
|
return bad_readline();
|
|
|
|
errno = 0;
|
|
/* XXX: Should the base argument of strtol() be explicitly set to 10?
|
|
XXX(avassalotti): Should this uses PyOS_strtol()? */
|
|
x = strtol(s, &endptr, 0);
|
|
|
|
if (errno || (*endptr != '\n' && *endptr != '\0')) {
|
|
/* Hm, maybe we've got something long. Let's try reading
|
|
* it as a Python int object. */
|
|
errno = 0;
|
|
/* XXX: Same thing about the base here. */
|
|
value = PyLong_FromString(s, NULL, 0);
|
|
if (value == NULL) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"could not convert string to int");
|
|
return -1;
|
|
}
|
|
}
|
|
else {
|
|
if (len == 3 && (x == 0 || x == 1)) {
|
|
if ((value = PyBool_FromLong(x)) == NULL)
|
|
return -1;
|
|
}
|
|
else {
|
|
if ((value = PyLong_FromLong(x)) == NULL)
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
PDATA_PUSH(self->stack, value, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_bool(UnpicklerObject *self, PyObject *boolean)
|
|
{
|
|
assert(boolean == Py_True || boolean == Py_False);
|
|
PDATA_APPEND(self->stack, boolean, -1);
|
|
return 0;
|
|
}
|
|
|
|
/* s contains x bytes of an unsigned little-endian integer. Return its value
|
|
* as a C Py_ssize_t, or -1 if it's higher than PY_SSIZE_T_MAX.
|
|
*/
|
|
static Py_ssize_t
|
|
calc_binsize(char *bytes, int nbytes)
|
|
{
|
|
unsigned char *s = (unsigned char *)bytes;
|
|
int i;
|
|
size_t x = 0;
|
|
|
|
if (nbytes > (int)sizeof(size_t)) {
|
|
/* Check for integer overflow. BINBYTES8 and BINUNICODE8 opcodes
|
|
* have 64-bit size that can't be represented on 32-bit platform.
|
|
*/
|
|
for (i = (int)sizeof(size_t); i < nbytes; i++) {
|
|
if (s[i])
|
|
return -1;
|
|
}
|
|
nbytes = (int)sizeof(size_t);
|
|
}
|
|
for (i = 0; i < nbytes; i++) {
|
|
x |= (size_t) s[i] << (8 * i);
|
|
}
|
|
|
|
if (x > PY_SSIZE_T_MAX)
|
|
return -1;
|
|
else
|
|
return (Py_ssize_t) x;
|
|
}
|
|
|
|
/* s contains x bytes of a little-endian integer. Return its value as a
|
|
* C int. Obscure: when x is 1 or 2, this is an unsigned little-endian
|
|
* int, but when x is 4 it's a signed one. This is a historical source
|
|
* of x-platform bugs.
|
|
*/
|
|
static long
|
|
calc_binint(char *bytes, int nbytes)
|
|
{
|
|
unsigned char *s = (unsigned char *)bytes;
|
|
Py_ssize_t i;
|
|
long x = 0;
|
|
|
|
for (i = 0; i < nbytes; i++) {
|
|
x |= (long)s[i] << (8 * i);
|
|
}
|
|
|
|
/* Unlike BININT1 and BININT2, BININT (more accurately BININT4)
|
|
* is signed, so on a box with longs bigger than 4 bytes we need
|
|
* to extend a BININT's sign bit to the full width.
|
|
*/
|
|
if (SIZEOF_LONG > 4 && nbytes == 4) {
|
|
x |= -(x & (1L << 31));
|
|
}
|
|
|
|
return x;
|
|
}
|
|
|
|
static int
|
|
load_binintx(UnpicklerObject *self, char *s, int size)
|
|
{
|
|
PyObject *value;
|
|
long x;
|
|
|
|
x = calc_binint(s, size);
|
|
|
|
if ((value = PyLong_FromLong(x)) == NULL)
|
|
return -1;
|
|
|
|
PDATA_PUSH(self->stack, value, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_binint(UnpicklerObject *self)
|
|
{
|
|
char *s;
|
|
|
|
if (_Unpickler_Read(self, &s, 4) < 0)
|
|
return -1;
|
|
|
|
return load_binintx(self, s, 4);
|
|
}
|
|
|
|
static int
|
|
load_binint1(UnpicklerObject *self)
|
|
{
|
|
char *s;
|
|
|
|
if (_Unpickler_Read(self, &s, 1) < 0)
|
|
return -1;
|
|
|
|
return load_binintx(self, s, 1);
|
|
}
|
|
|
|
static int
|
|
load_binint2(UnpicklerObject *self)
|
|
{
|
|
char *s;
|
|
|
|
if (_Unpickler_Read(self, &s, 2) < 0)
|
|
return -1;
|
|
|
|
return load_binintx(self, s, 2);
|
|
}
|
|
|
|
static int
|
|
load_long(UnpicklerObject *self)
|
|
{
|
|
PyObject *value;
|
|
char *s = NULL;
|
|
Py_ssize_t len;
|
|
|
|
if ((len = _Unpickler_Readline(self, &s)) < 0)
|
|
return -1;
|
|
if (len < 2)
|
|
return bad_readline();
|
|
|
|
/* s[len-2] will usually be 'L' (and s[len-1] is '\n'); we need to remove
|
|
the 'L' before calling PyLong_FromString. In order to maintain
|
|
compatibility with Python 3.0.0, we don't actually *require*
|
|
the 'L' to be present. */
|
|
if (s[len-2] == 'L')
|
|
s[len-2] = '\0';
|
|
/* XXX: Should the base argument explicitly set to 10? */
|
|
value = PyLong_FromString(s, NULL, 0);
|
|
if (value == NULL)
|
|
return -1;
|
|
|
|
PDATA_PUSH(self->stack, value, -1);
|
|
return 0;
|
|
}
|
|
|
|
/* 'size' bytes contain the # of bytes of little-endian 256's-complement
|
|
* data following.
|
|
*/
|
|
static int
|
|
load_counted_long(UnpicklerObject *self, int size)
|
|
{
|
|
PyObject *value;
|
|
char *nbytes;
|
|
char *pdata;
|
|
|
|
assert(size == 1 || size == 4);
|
|
if (_Unpickler_Read(self, &nbytes, size) < 0)
|
|
return -1;
|
|
|
|
size = calc_binint(nbytes, size);
|
|
if (size < 0) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
/* Corrupt or hostile pickle -- we never write one like this */
|
|
PyErr_SetString(st->UnpicklingError,
|
|
"LONG pickle has negative byte count");
|
|
return -1;
|
|
}
|
|
|
|
if (size == 0)
|
|
value = PyLong_FromLong(0L);
|
|
else {
|
|
/* Read the raw little-endian bytes and convert. */
|
|
if (_Unpickler_Read(self, &pdata, size) < 0)
|
|
return -1;
|
|
value = _PyLong_FromByteArray((unsigned char *)pdata, (size_t)size,
|
|
1 /* little endian */ , 1 /* signed */ );
|
|
}
|
|
if (value == NULL)
|
|
return -1;
|
|
PDATA_PUSH(self->stack, value, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_float(UnpicklerObject *self)
|
|
{
|
|
PyObject *value;
|
|
char *endptr, *s;
|
|
Py_ssize_t len;
|
|
double d;
|
|
|
|
if ((len = _Unpickler_Readline(self, &s)) < 0)
|
|
return -1;
|
|
if (len < 2)
|
|
return bad_readline();
|
|
|
|
errno = 0;
|
|
d = PyOS_string_to_double(s, &endptr, PyExc_OverflowError);
|
|
if (d == -1.0 && PyErr_Occurred())
|
|
return -1;
|
|
if ((endptr[0] != '\n') && (endptr[0] != '\0')) {
|
|
PyErr_SetString(PyExc_ValueError, "could not convert string to float");
|
|
return -1;
|
|
}
|
|
value = PyFloat_FromDouble(d);
|
|
if (value == NULL)
|
|
return -1;
|
|
|
|
PDATA_PUSH(self->stack, value, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_binfloat(UnpicklerObject *self)
|
|
{
|
|
PyObject *value;
|
|
double x;
|
|
char *s;
|
|
|
|
if (_Unpickler_Read(self, &s, 8) < 0)
|
|
return -1;
|
|
|
|
x = _PyFloat_Unpack8((unsigned char *)s, 0);
|
|
if (x == -1.0 && PyErr_Occurred())
|
|
return -1;
|
|
|
|
if ((value = PyFloat_FromDouble(x)) == NULL)
|
|
return -1;
|
|
|
|
PDATA_PUSH(self->stack, value, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_string(UnpicklerObject *self)
|
|
{
|
|
PyObject *bytes;
|
|
PyObject *obj;
|
|
Py_ssize_t len;
|
|
char *s, *p;
|
|
|
|
if ((len = _Unpickler_Readline(self, &s)) < 0)
|
|
return -1;
|
|
/* Strip the newline */
|
|
len--;
|
|
/* Strip outermost quotes */
|
|
if (len >= 2 && s[0] == s[len - 1] && (s[0] == '\'' || s[0] == '"')) {
|
|
p = s + 1;
|
|
len -= 2;
|
|
}
|
|
else {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->UnpicklingError,
|
|
"the STRING opcode argument must be quoted");
|
|
return -1;
|
|
}
|
|
assert(len >= 0);
|
|
|
|
/* Use the PyBytes API to decode the string, since that is what is used
|
|
to encode, and then coerce the result to Unicode. */
|
|
bytes = PyBytes_DecodeEscape(p, len, NULL, 0, NULL);
|
|
if (bytes == NULL)
|
|
return -1;
|
|
|
|
/* Leave the Python 2.x strings as bytes if the *encoding* given to the
|
|
Unpickler was 'bytes'. Otherwise, convert them to unicode. */
|
|
if (strcmp(self->encoding, "bytes") == 0) {
|
|
obj = bytes;
|
|
}
|
|
else {
|
|
obj = PyUnicode_FromEncodedObject(bytes, self->encoding, self->errors);
|
|
Py_DECREF(bytes);
|
|
if (obj == NULL) {
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
PDATA_PUSH(self->stack, obj, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_counted_binstring(UnpicklerObject *self, int nbytes)
|
|
{
|
|
PyObject *obj;
|
|
Py_ssize_t size;
|
|
char *s;
|
|
|
|
if (_Unpickler_Read(self, &s, nbytes) < 0)
|
|
return -1;
|
|
|
|
size = calc_binsize(s, nbytes);
|
|
if (size < 0) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_Format(st->UnpicklingError,
|
|
"BINSTRING exceeds system's maximum size of %zd bytes",
|
|
PY_SSIZE_T_MAX);
|
|
return -1;
|
|
}
|
|
|
|
if (_Unpickler_Read(self, &s, size) < 0)
|
|
return -1;
|
|
|
|
/* Convert Python 2.x strings to bytes if the *encoding* given to the
|
|
Unpickler was 'bytes'. Otherwise, convert them to unicode. */
|
|
if (strcmp(self->encoding, "bytes") == 0) {
|
|
obj = PyBytes_FromStringAndSize(s, size);
|
|
}
|
|
else {
|
|
obj = PyUnicode_Decode(s, size, self->encoding, self->errors);
|
|
}
|
|
if (obj == NULL) {
|
|
return -1;
|
|
}
|
|
|
|
PDATA_PUSH(self->stack, obj, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_counted_binbytes(UnpicklerObject *self, int nbytes)
|
|
{
|
|
PyObject *bytes;
|
|
Py_ssize_t size;
|
|
char *s;
|
|
|
|
if (_Unpickler_Read(self, &s, nbytes) < 0)
|
|
return -1;
|
|
|
|
size = calc_binsize(s, nbytes);
|
|
if (size < 0) {
|
|
PyErr_Format(PyExc_OverflowError,
|
|
"BINBYTES exceeds system's maximum size of %zd bytes",
|
|
PY_SSIZE_T_MAX);
|
|
return -1;
|
|
}
|
|
|
|
bytes = PyBytes_FromStringAndSize(NULL, size);
|
|
if (bytes == NULL)
|
|
return -1;
|
|
if (_Unpickler_ReadInto(self, PyBytes_AS_STRING(bytes), size) < 0) {
|
|
Py_DECREF(bytes);
|
|
return -1;
|
|
}
|
|
|
|
PDATA_PUSH(self->stack, bytes, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_counted_bytearray(UnpicklerObject *self)
|
|
{
|
|
PyObject *bytearray;
|
|
Py_ssize_t size;
|
|
char *s;
|
|
|
|
if (_Unpickler_Read(self, &s, 8) < 0) {
|
|
return -1;
|
|
}
|
|
|
|
size = calc_binsize(s, 8);
|
|
if (size < 0) {
|
|
PyErr_Format(PyExc_OverflowError,
|
|
"BYTEARRAY8 exceeds system's maximum size of %zd bytes",
|
|
PY_SSIZE_T_MAX);
|
|
return -1;
|
|
}
|
|
|
|
bytearray = PyByteArray_FromStringAndSize(NULL, size);
|
|
if (bytearray == NULL) {
|
|
return -1;
|
|
}
|
|
if (_Unpickler_ReadInto(self, PyByteArray_AS_STRING(bytearray), size) < 0) {
|
|
Py_DECREF(bytearray);
|
|
return -1;
|
|
}
|
|
|
|
PDATA_PUSH(self->stack, bytearray, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_next_buffer(UnpicklerObject *self)
|
|
{
|
|
if (self->buffers == NULL) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->UnpicklingError,
|
|
"pickle stream refers to out-of-band data "
|
|
"but no *buffers* argument was given");
|
|
return -1;
|
|
}
|
|
PyObject *buf = PyIter_Next(self->buffers);
|
|
if (buf == NULL) {
|
|
if (!PyErr_Occurred()) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->UnpicklingError,
|
|
"not enough out-of-band buffers");
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
PDATA_PUSH(self->stack, buf, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_readonly_buffer(UnpicklerObject *self)
|
|
{
|
|
Py_ssize_t len = Py_SIZE(self->stack);
|
|
if (len <= self->stack->fence) {
|
|
return Pdata_stack_underflow(self->stack);
|
|
}
|
|
|
|
PyObject *obj = self->stack->data[len - 1];
|
|
PyObject *view = PyMemoryView_FromObject(obj);
|
|
if (view == NULL) {
|
|
return -1;
|
|
}
|
|
if (!PyMemoryView_GET_BUFFER(view)->readonly) {
|
|
/* Original object is writable */
|
|
PyMemoryView_GET_BUFFER(view)->readonly = 1;
|
|
self->stack->data[len - 1] = view;
|
|
Py_DECREF(obj);
|
|
}
|
|
else {
|
|
/* Original object is read-only, no need to replace it */
|
|
Py_DECREF(view);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_unicode(UnpicklerObject *self)
|
|
{
|
|
PyObject *str;
|
|
Py_ssize_t len;
|
|
char *s = NULL;
|
|
|
|
if ((len = _Unpickler_Readline(self, &s)) < 0)
|
|
return -1;
|
|
if (len < 1)
|
|
return bad_readline();
|
|
|
|
str = PyUnicode_DecodeRawUnicodeEscape(s, len - 1, NULL);
|
|
if (str == NULL)
|
|
return -1;
|
|
|
|
PDATA_PUSH(self->stack, str, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_counted_binunicode(UnpicklerObject *self, int nbytes)
|
|
{
|
|
PyObject *str;
|
|
Py_ssize_t size;
|
|
char *s;
|
|
|
|
if (_Unpickler_Read(self, &s, nbytes) < 0)
|
|
return -1;
|
|
|
|
size = calc_binsize(s, nbytes);
|
|
if (size < 0) {
|
|
PyErr_Format(PyExc_OverflowError,
|
|
"BINUNICODE exceeds system's maximum size of %zd bytes",
|
|
PY_SSIZE_T_MAX);
|
|
return -1;
|
|
}
|
|
|
|
if (_Unpickler_Read(self, &s, size) < 0)
|
|
return -1;
|
|
|
|
str = PyUnicode_DecodeUTF8(s, size, "surrogatepass");
|
|
if (str == NULL)
|
|
return -1;
|
|
|
|
PDATA_PUSH(self->stack, str, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_counted_tuple(UnpicklerObject *self, Py_ssize_t len)
|
|
{
|
|
PyObject *tuple;
|
|
|
|
if (Py_SIZE(self->stack) < len)
|
|
return Pdata_stack_underflow(self->stack);
|
|
|
|
tuple = Pdata_poptuple(self->stack, Py_SIZE(self->stack) - len);
|
|
if (tuple == NULL)
|
|
return -1;
|
|
PDATA_PUSH(self->stack, tuple, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_tuple(UnpicklerObject *self)
|
|
{
|
|
Py_ssize_t i;
|
|
|
|
if ((i = marker(self)) < 0)
|
|
return -1;
|
|
|
|
return load_counted_tuple(self, Py_SIZE(self->stack) - i);
|
|
}
|
|
|
|
static int
|
|
load_empty_list(UnpicklerObject *self)
|
|
{
|
|
PyObject *list;
|
|
|
|
if ((list = PyList_New(0)) == NULL)
|
|
return -1;
|
|
PDATA_PUSH(self->stack, list, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_empty_dict(UnpicklerObject *self)
|
|
{
|
|
PyObject *dict;
|
|
|
|
if ((dict = PyDict_New()) == NULL)
|
|
return -1;
|
|
PDATA_PUSH(self->stack, dict, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_empty_set(UnpicklerObject *self)
|
|
{
|
|
PyObject *set;
|
|
|
|
if ((set = PySet_New(NULL)) == NULL)
|
|
return -1;
|
|
PDATA_PUSH(self->stack, set, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_list(UnpicklerObject *self)
|
|
{
|
|
PyObject *list;
|
|
Py_ssize_t i;
|
|
|
|
if ((i = marker(self)) < 0)
|
|
return -1;
|
|
|
|
list = Pdata_poplist(self->stack, i);
|
|
if (list == NULL)
|
|
return -1;
|
|
PDATA_PUSH(self->stack, list, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_dict(UnpicklerObject *self)
|
|
{
|
|
PyObject *dict, *key, *value;
|
|
Py_ssize_t i, j, k;
|
|
|
|
if ((i = marker(self)) < 0)
|
|
return -1;
|
|
j = Py_SIZE(self->stack);
|
|
|
|
if ((dict = PyDict_New()) == NULL)
|
|
return -1;
|
|
|
|
if ((j - i) % 2 != 0) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->UnpicklingError, "odd number of items for DICT");
|
|
Py_DECREF(dict);
|
|
return -1;
|
|
}
|
|
|
|
for (k = i + 1; k < j; k += 2) {
|
|
key = self->stack->data[k - 1];
|
|
value = self->stack->data[k];
|
|
if (PyDict_SetItem(dict, key, value) < 0) {
|
|
Py_DECREF(dict);
|
|
return -1;
|
|
}
|
|
}
|
|
Pdata_clear(self->stack, i);
|
|
PDATA_PUSH(self->stack, dict, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_frozenset(UnpicklerObject *self)
|
|
{
|
|
PyObject *items;
|
|
PyObject *frozenset;
|
|
Py_ssize_t i;
|
|
|
|
if ((i = marker(self)) < 0)
|
|
return -1;
|
|
|
|
items = Pdata_poptuple(self->stack, i);
|
|
if (items == NULL)
|
|
return -1;
|
|
|
|
frozenset = PyFrozenSet_New(items);
|
|
Py_DECREF(items);
|
|
if (frozenset == NULL)
|
|
return -1;
|
|
|
|
PDATA_PUSH(self->stack, frozenset, -1);
|
|
return 0;
|
|
}
|
|
|
|
static PyObject *
|
|
instantiate(PyObject *cls, PyObject *args)
|
|
{
|
|
/* Caller must assure args are a tuple. Normally, args come from
|
|
Pdata_poptuple which packs objects from the top of the stack
|
|
into a newly created tuple. */
|
|
assert(PyTuple_Check(args));
|
|
if (!PyTuple_GET_SIZE(args) && PyType_Check(cls)) {
|
|
_Py_IDENTIFIER(__getinitargs__);
|
|
_Py_IDENTIFIER(__new__);
|
|
PyObject *func;
|
|
if (_PyObject_LookupAttrId(cls, &PyId___getinitargs__, &func) < 0) {
|
|
return NULL;
|
|
}
|
|
if (func == NULL) {
|
|
return _PyObject_CallMethodIdOneArg(cls, &PyId___new__, cls);
|
|
}
|
|
Py_DECREF(func);
|
|
}
|
|
return PyObject_CallObject(cls, args);
|
|
}
|
|
|
|
static int
|
|
load_obj(UnpicklerObject *self)
|
|
{
|
|
PyObject *cls, *args, *obj = NULL;
|
|
Py_ssize_t i;
|
|
|
|
if ((i = marker(self)) < 0)
|
|
return -1;
|
|
|
|
if (Py_SIZE(self->stack) - i < 1)
|
|
return Pdata_stack_underflow(self->stack);
|
|
|
|
args = Pdata_poptuple(self->stack, i + 1);
|
|
if (args == NULL)
|
|
return -1;
|
|
|
|
PDATA_POP(self->stack, cls);
|
|
if (cls) {
|
|
obj = instantiate(cls, args);
|
|
Py_DECREF(cls);
|
|
}
|
|
Py_DECREF(args);
|
|
if (obj == NULL)
|
|
return -1;
|
|
|
|
PDATA_PUSH(self->stack, obj, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_inst(UnpicklerObject *self)
|
|
{
|
|
PyObject *cls = NULL;
|
|
PyObject *args = NULL;
|
|
PyObject *obj = NULL;
|
|
PyObject *module_name;
|
|
PyObject *class_name;
|
|
Py_ssize_t len;
|
|
Py_ssize_t i;
|
|
char *s;
|
|
|
|
if ((i = marker(self)) < 0)
|
|
return -1;
|
|
if ((len = _Unpickler_Readline(self, &s)) < 0)
|
|
return -1;
|
|
if (len < 2)
|
|
return bad_readline();
|
|
|
|
/* Here it is safe to use PyUnicode_DecodeASCII(), even though non-ASCII
|
|
identifiers are permitted in Python 3.0, since the INST opcode is only
|
|
supported by older protocols on Python 2.x. */
|
|
module_name = PyUnicode_DecodeASCII(s, len - 1, "strict");
|
|
if (module_name == NULL)
|
|
return -1;
|
|
|
|
if ((len = _Unpickler_Readline(self, &s)) >= 0) {
|
|
if (len < 2) {
|
|
Py_DECREF(module_name);
|
|
return bad_readline();
|
|
}
|
|
class_name = PyUnicode_DecodeASCII(s, len - 1, "strict");
|
|
if (class_name != NULL) {
|
|
cls = find_class(self, module_name, class_name);
|
|
Py_DECREF(class_name);
|
|
}
|
|
}
|
|
Py_DECREF(module_name);
|
|
|
|
if (cls == NULL)
|
|
return -1;
|
|
|
|
if ((args = Pdata_poptuple(self->stack, i)) != NULL) {
|
|
obj = instantiate(cls, args);
|
|
Py_DECREF(args);
|
|
}
|
|
Py_DECREF(cls);
|
|
|
|
if (obj == NULL)
|
|
return -1;
|
|
|
|
PDATA_PUSH(self->stack, obj, -1);
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
newobj_unpickling_error(const char * msg, int use_kwargs, PyObject *arg)
|
|
{
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_Format(st->UnpicklingError, msg,
|
|
use_kwargs ? "NEWOBJ_EX" : "NEWOBJ",
|
|
Py_TYPE(arg)->tp_name);
|
|
}
|
|
|
|
static int
|
|
load_newobj(UnpicklerObject *self, int use_kwargs)
|
|
{
|
|
PyObject *cls, *args, *kwargs = NULL;
|
|
PyObject *obj;
|
|
|
|
/* Stack is ... cls args [kwargs], and we want to call
|
|
* cls.__new__(cls, *args, **kwargs).
|
|
*/
|
|
if (use_kwargs) {
|
|
PDATA_POP(self->stack, kwargs);
|
|
if (kwargs == NULL) {
|
|
return -1;
|
|
}
|
|
}
|
|
PDATA_POP(self->stack, args);
|
|
if (args == NULL) {
|
|
Py_XDECREF(kwargs);
|
|
return -1;
|
|
}
|
|
PDATA_POP(self->stack, cls);
|
|
if (cls == NULL) {
|
|
Py_XDECREF(kwargs);
|
|
Py_DECREF(args);
|
|
return -1;
|
|
}
|
|
|
|
if (!PyType_Check(cls)) {
|
|
newobj_unpickling_error("%s class argument must be a type, not %.200s",
|
|
use_kwargs, cls);
|
|
goto error;
|
|
}
|
|
if (((PyTypeObject *)cls)->tp_new == NULL) {
|
|
newobj_unpickling_error("%s class argument '%.200s' doesn't have __new__",
|
|
use_kwargs, cls);
|
|
goto error;
|
|
}
|
|
if (!PyTuple_Check(args)) {
|
|
newobj_unpickling_error("%s args argument must be a tuple, not %.200s",
|
|
use_kwargs, args);
|
|
goto error;
|
|
}
|
|
if (use_kwargs && !PyDict_Check(kwargs)) {
|
|
newobj_unpickling_error("%s kwargs argument must be a dict, not %.200s",
|
|
use_kwargs, kwargs);
|
|
goto error;
|
|
}
|
|
|
|
obj = ((PyTypeObject *)cls)->tp_new((PyTypeObject *)cls, args, kwargs);
|
|
if (obj == NULL) {
|
|
goto error;
|
|
}
|
|
Py_XDECREF(kwargs);
|
|
Py_DECREF(args);
|
|
Py_DECREF(cls);
|
|
PDATA_PUSH(self->stack, obj, -1);
|
|
return 0;
|
|
|
|
error:
|
|
Py_XDECREF(kwargs);
|
|
Py_DECREF(args);
|
|
Py_DECREF(cls);
|
|
return -1;
|
|
}
|
|
|
|
static int
|
|
load_global(UnpicklerObject *self)
|
|
{
|
|
PyObject *global = NULL;
|
|
PyObject *module_name;
|
|
PyObject *global_name;
|
|
Py_ssize_t len;
|
|
char *s;
|
|
|
|
if ((len = _Unpickler_Readline(self, &s)) < 0)
|
|
return -1;
|
|
if (len < 2)
|
|
return bad_readline();
|
|
module_name = PyUnicode_DecodeUTF8(s, len - 1, "strict");
|
|
if (!module_name)
|
|
return -1;
|
|
|
|
if ((len = _Unpickler_Readline(self, &s)) >= 0) {
|
|
if (len < 2) {
|
|
Py_DECREF(module_name);
|
|
return bad_readline();
|
|
}
|
|
global_name = PyUnicode_DecodeUTF8(s, len - 1, "strict");
|
|
if (global_name) {
|
|
global = find_class(self, module_name, global_name);
|
|
Py_DECREF(global_name);
|
|
}
|
|
}
|
|
Py_DECREF(module_name);
|
|
|
|
if (global == NULL)
|
|
return -1;
|
|
PDATA_PUSH(self->stack, global, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_stack_global(UnpicklerObject *self)
|
|
{
|
|
PyObject *global;
|
|
PyObject *module_name;
|
|
PyObject *global_name;
|
|
|
|
PDATA_POP(self->stack, global_name);
|
|
PDATA_POP(self->stack, module_name);
|
|
if (module_name == NULL || !PyUnicode_CheckExact(module_name) ||
|
|
global_name == NULL || !PyUnicode_CheckExact(global_name)) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->UnpicklingError, "STACK_GLOBAL requires str");
|
|
Py_XDECREF(global_name);
|
|
Py_XDECREF(module_name);
|
|
return -1;
|
|
}
|
|
global = find_class(self, module_name, global_name);
|
|
Py_DECREF(global_name);
|
|
Py_DECREF(module_name);
|
|
if (global == NULL)
|
|
return -1;
|
|
PDATA_PUSH(self->stack, global, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_persid(UnpicklerObject *self)
|
|
{
|
|
PyObject *pid, *obj;
|
|
Py_ssize_t len;
|
|
char *s;
|
|
|
|
if (self->pers_func) {
|
|
if ((len = _Unpickler_Readline(self, &s)) < 0)
|
|
return -1;
|
|
if (len < 1)
|
|
return bad_readline();
|
|
|
|
pid = PyUnicode_DecodeASCII(s, len - 1, "strict");
|
|
if (pid == NULL) {
|
|
if (PyErr_ExceptionMatches(PyExc_UnicodeDecodeError)) {
|
|
PyErr_SetString(_Pickle_GetGlobalState()->UnpicklingError,
|
|
"persistent IDs in protocol 0 must be "
|
|
"ASCII strings");
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
obj = call_method(self->pers_func, self->pers_func_self, pid);
|
|
Py_DECREF(pid);
|
|
if (obj == NULL)
|
|
return -1;
|
|
|
|
PDATA_PUSH(self->stack, obj, -1);
|
|
return 0;
|
|
}
|
|
else {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->UnpicklingError,
|
|
"A load persistent id instruction was encountered,\n"
|
|
"but no persistent_load function was specified.");
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
static int
|
|
load_binpersid(UnpicklerObject *self)
|
|
{
|
|
PyObject *pid, *obj;
|
|
|
|
if (self->pers_func) {
|
|
PDATA_POP(self->stack, pid);
|
|
if (pid == NULL)
|
|
return -1;
|
|
|
|
obj = call_method(self->pers_func, self->pers_func_self, pid);
|
|
Py_DECREF(pid);
|
|
if (obj == NULL)
|
|
return -1;
|
|
|
|
PDATA_PUSH(self->stack, obj, -1);
|
|
return 0;
|
|
}
|
|
else {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->UnpicklingError,
|
|
"A load persistent id instruction was encountered,\n"
|
|
"but no persistent_load function was specified.");
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
static int
|
|
load_pop(UnpicklerObject *self)
|
|
{
|
|
Py_ssize_t len = Py_SIZE(self->stack);
|
|
|
|
/* Note that we split the (pickle.py) stack into two stacks,
|
|
* an object stack and a mark stack. We have to be clever and
|
|
* pop the right one. We do this by looking at the top of the
|
|
* mark stack first, and only signalling a stack underflow if
|
|
* the object stack is empty and the mark stack doesn't match
|
|
* our expectations.
|
|
*/
|
|
if (self->num_marks > 0 && self->marks[self->num_marks - 1] == len) {
|
|
self->num_marks--;
|
|
self->stack->mark_set = self->num_marks != 0;
|
|
self->stack->fence = self->num_marks ?
|
|
self->marks[self->num_marks - 1] : 0;
|
|
} else if (len <= self->stack->fence)
|
|
return Pdata_stack_underflow(self->stack);
|
|
else {
|
|
len--;
|
|
Py_DECREF(self->stack->data[len]);
|
|
Py_SET_SIZE(self->stack, len);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_pop_mark(UnpicklerObject *self)
|
|
{
|
|
Py_ssize_t i;
|
|
|
|
if ((i = marker(self)) < 0)
|
|
return -1;
|
|
|
|
Pdata_clear(self->stack, i);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_dup(UnpicklerObject *self)
|
|
{
|
|
PyObject *last;
|
|
Py_ssize_t len = Py_SIZE(self->stack);
|
|
|
|
if (len <= self->stack->fence)
|
|
return Pdata_stack_underflow(self->stack);
|
|
last = self->stack->data[len - 1];
|
|
PDATA_APPEND(self->stack, last, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_get(UnpicklerObject *self)
|
|
{
|
|
PyObject *key, *value;
|
|
Py_ssize_t idx;
|
|
Py_ssize_t len;
|
|
char *s;
|
|
|
|
if ((len = _Unpickler_Readline(self, &s)) < 0)
|
|
return -1;
|
|
if (len < 2)
|
|
return bad_readline();
|
|
|
|
key = PyLong_FromString(s, NULL, 10);
|
|
if (key == NULL)
|
|
return -1;
|
|
idx = PyLong_AsSsize_t(key);
|
|
if (idx == -1 && PyErr_Occurred()) {
|
|
Py_DECREF(key);
|
|
return -1;
|
|
}
|
|
|
|
value = _Unpickler_MemoGet(self, idx);
|
|
if (value == NULL) {
|
|
if (!PyErr_Occurred()) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_Format(st->UnpicklingError, "Memo value not found at index %ld", idx);
|
|
}
|
|
Py_DECREF(key);
|
|
return -1;
|
|
}
|
|
Py_DECREF(key);
|
|
|
|
PDATA_APPEND(self->stack, value, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_binget(UnpicklerObject *self)
|
|
{
|
|
PyObject *value;
|
|
Py_ssize_t idx;
|
|
char *s;
|
|
|
|
if (_Unpickler_Read(self, &s, 1) < 0)
|
|
return -1;
|
|
|
|
idx = Py_CHARMASK(s[0]);
|
|
|
|
value = _Unpickler_MemoGet(self, idx);
|
|
if (value == NULL) {
|
|
PyObject *key = PyLong_FromSsize_t(idx);
|
|
if (key != NULL) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_Format(st->UnpicklingError, "Memo value not found at index %ld", idx);
|
|
Py_DECREF(key);
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
PDATA_APPEND(self->stack, value, -1);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_long_binget(UnpicklerObject *self)
|
|
{
|
|
PyObject *value;
|
|
Py_ssize_t idx;
|
|
char *s;
|
|
|
|
if (_Unpickler_Read(self, &s, 4) < 0)
|
|
return -1;
|
|
|
|
idx = calc_binsize(s, 4);
|
|
|
|
value = _Unpickler_MemoGet(self, idx);
|
|
if (value == NULL) {
|
|
PyObject *key = PyLong_FromSsize_t(idx);
|
|
if (key != NULL) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_Format(st->UnpicklingError, "Memo value not found at index %ld", idx);
|
|
Py_DECREF(key);
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
PDATA_APPEND(self->stack, value, -1);
|
|
return 0;
|
|
}
|
|
|
|
/* Push an object from the extension registry (EXT[124]). nbytes is
|
|
* the number of bytes following the opcode, holding the index (code) value.
|
|
*/
|
|
static int
|
|
load_extension(UnpicklerObject *self, int nbytes)
|
|
{
|
|
char *codebytes; /* the nbytes bytes after the opcode */
|
|
long code; /* calc_binint returns long */
|
|
PyObject *py_code; /* code as a Python int */
|
|
PyObject *obj; /* the object to push */
|
|
PyObject *pair; /* (module_name, class_name) */
|
|
PyObject *module_name, *class_name;
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
|
|
assert(nbytes == 1 || nbytes == 2 || nbytes == 4);
|
|
if (_Unpickler_Read(self, &codebytes, nbytes) < 0)
|
|
return -1;
|
|
code = calc_binint(codebytes, nbytes);
|
|
if (code <= 0) { /* note that 0 is forbidden */
|
|
/* Corrupt or hostile pickle. */
|
|
PyErr_SetString(st->UnpicklingError, "EXT specifies code <= 0");
|
|
return -1;
|
|
}
|
|
|
|
/* Look for the code in the cache. */
|
|
py_code = PyLong_FromLong(code);
|
|
if (py_code == NULL)
|
|
return -1;
|
|
obj = PyDict_GetItemWithError(st->extension_cache, py_code);
|
|
if (obj != NULL) {
|
|
/* Bingo. */
|
|
Py_DECREF(py_code);
|
|
PDATA_APPEND(self->stack, obj, -1);
|
|
return 0;
|
|
}
|
|
if (PyErr_Occurred()) {
|
|
Py_DECREF(py_code);
|
|
return -1;
|
|
}
|
|
|
|
/* Look up the (module_name, class_name) pair. */
|
|
pair = PyDict_GetItemWithError(st->inverted_registry, py_code);
|
|
if (pair == NULL) {
|
|
Py_DECREF(py_code);
|
|
if (!PyErr_Occurred()) {
|
|
PyErr_Format(PyExc_ValueError, "unregistered extension "
|
|
"code %ld", code);
|
|
}
|
|
return -1;
|
|
}
|
|
/* Since the extension registry is manipulable via Python code,
|
|
* confirm that pair is really a 2-tuple of strings.
|
|
*/
|
|
if (!PyTuple_Check(pair) || PyTuple_Size(pair) != 2) {
|
|
goto error;
|
|
}
|
|
|
|
module_name = PyTuple_GET_ITEM(pair, 0);
|
|
if (!PyUnicode_Check(module_name)) {
|
|
goto error;
|
|
}
|
|
|
|
class_name = PyTuple_GET_ITEM(pair, 1);
|
|
if (!PyUnicode_Check(class_name)) {
|
|
goto error;
|
|
}
|
|
|
|
/* Load the object. */
|
|
obj = find_class(self, module_name, class_name);
|
|
if (obj == NULL) {
|
|
Py_DECREF(py_code);
|
|
return -1;
|
|
}
|
|
/* Cache code -> obj. */
|
|
code = PyDict_SetItem(st->extension_cache, py_code, obj);
|
|
Py_DECREF(py_code);
|
|
if (code < 0) {
|
|
Py_DECREF(obj);
|
|
return -1;
|
|
}
|
|
PDATA_PUSH(self->stack, obj, -1);
|
|
return 0;
|
|
|
|
error:
|
|
Py_DECREF(py_code);
|
|
PyErr_Format(PyExc_ValueError, "_inverted_registry[%ld] "
|
|
"isn't a 2-tuple of strings", code);
|
|
return -1;
|
|
}
|
|
|
|
static int
|
|
load_put(UnpicklerObject *self)
|
|
{
|
|
PyObject *key, *value;
|
|
Py_ssize_t idx;
|
|
Py_ssize_t len;
|
|
char *s = NULL;
|
|
|
|
if ((len = _Unpickler_Readline(self, &s)) < 0)
|
|
return -1;
|
|
if (len < 2)
|
|
return bad_readline();
|
|
if (Py_SIZE(self->stack) <= self->stack->fence)
|
|
return Pdata_stack_underflow(self->stack);
|
|
value = self->stack->data[Py_SIZE(self->stack) - 1];
|
|
|
|
key = PyLong_FromString(s, NULL, 10);
|
|
if (key == NULL)
|
|
return -1;
|
|
idx = PyLong_AsSsize_t(key);
|
|
Py_DECREF(key);
|
|
if (idx < 0) {
|
|
if (!PyErr_Occurred())
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"negative PUT argument");
|
|
return -1;
|
|
}
|
|
|
|
return _Unpickler_MemoPut(self, idx, value);
|
|
}
|
|
|
|
static int
|
|
load_binput(UnpicklerObject *self)
|
|
{
|
|
PyObject *value;
|
|
Py_ssize_t idx;
|
|
char *s;
|
|
|
|
if (_Unpickler_Read(self, &s, 1) < 0)
|
|
return -1;
|
|
|
|
if (Py_SIZE(self->stack) <= self->stack->fence)
|
|
return Pdata_stack_underflow(self->stack);
|
|
value = self->stack->data[Py_SIZE(self->stack) - 1];
|
|
|
|
idx = Py_CHARMASK(s[0]);
|
|
|
|
return _Unpickler_MemoPut(self, idx, value);
|
|
}
|
|
|
|
static int
|
|
load_long_binput(UnpicklerObject *self)
|
|
{
|
|
PyObject *value;
|
|
Py_ssize_t idx;
|
|
char *s;
|
|
|
|
if (_Unpickler_Read(self, &s, 4) < 0)
|
|
return -1;
|
|
|
|
if (Py_SIZE(self->stack) <= self->stack->fence)
|
|
return Pdata_stack_underflow(self->stack);
|
|
value = self->stack->data[Py_SIZE(self->stack) - 1];
|
|
|
|
idx = calc_binsize(s, 4);
|
|
if (idx < 0) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"negative LONG_BINPUT argument");
|
|
return -1;
|
|
}
|
|
|
|
return _Unpickler_MemoPut(self, idx, value);
|
|
}
|
|
|
|
static int
|
|
load_memoize(UnpicklerObject *self)
|
|
{
|
|
PyObject *value;
|
|
|
|
if (Py_SIZE(self->stack) <= self->stack->fence)
|
|
return Pdata_stack_underflow(self->stack);
|
|
value = self->stack->data[Py_SIZE(self->stack) - 1];
|
|
|
|
return _Unpickler_MemoPut(self, self->memo_len, value);
|
|
}
|
|
|
|
static int
|
|
do_append(UnpicklerObject *self, Py_ssize_t x)
|
|
{
|
|
PyObject *value;
|
|
PyObject *slice;
|
|
PyObject *list;
|
|
PyObject *result;
|
|
Py_ssize_t len, i;
|
|
|
|
len = Py_SIZE(self->stack);
|
|
if (x > len || x <= self->stack->fence)
|
|
return Pdata_stack_underflow(self->stack);
|
|
if (len == x) /* nothing to do */
|
|
return 0;
|
|
|
|
list = self->stack->data[x - 1];
|
|
|
|
if (PyList_CheckExact(list)) {
|
|
Py_ssize_t list_len;
|
|
int ret;
|
|
|
|
slice = Pdata_poplist(self->stack, x);
|
|
if (!slice)
|
|
return -1;
|
|
list_len = PyList_GET_SIZE(list);
|
|
ret = PyList_SetSlice(list, list_len, list_len, slice);
|
|
Py_DECREF(slice);
|
|
return ret;
|
|
}
|
|
else {
|
|
PyObject *extend_func;
|
|
_Py_IDENTIFIER(extend);
|
|
|
|
if (_PyObject_LookupAttrId(list, &PyId_extend, &extend_func) < 0) {
|
|
return -1;
|
|
}
|
|
if (extend_func != NULL) {
|
|
slice = Pdata_poplist(self->stack, x);
|
|
if (!slice) {
|
|
Py_DECREF(extend_func);
|
|
return -1;
|
|
}
|
|
result = _Pickle_FastCall(extend_func, slice);
|
|
Py_DECREF(extend_func);
|
|
if (result == NULL)
|
|
return -1;
|
|
Py_DECREF(result);
|
|
}
|
|
else {
|
|
PyObject *append_func;
|
|
_Py_IDENTIFIER(append);
|
|
|
|
/* Even if the PEP 307 requires extend() and append() methods,
|
|
fall back on append() if the object has no extend() method
|
|
for backward compatibility. */
|
|
append_func = _PyObject_GetAttrId(list, &PyId_append);
|
|
if (append_func == NULL)
|
|
return -1;
|
|
for (i = x; i < len; i++) {
|
|
value = self->stack->data[i];
|
|
result = _Pickle_FastCall(append_func, value);
|
|
if (result == NULL) {
|
|
Pdata_clear(self->stack, i + 1);
|
|
Py_SET_SIZE(self->stack, x);
|
|
Py_DECREF(append_func);
|
|
return -1;
|
|
}
|
|
Py_DECREF(result);
|
|
}
|
|
Py_SET_SIZE(self->stack, x);
|
|
Py_DECREF(append_func);
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_append(UnpicklerObject *self)
|
|
{
|
|
if (Py_SIZE(self->stack) - 1 <= self->stack->fence)
|
|
return Pdata_stack_underflow(self->stack);
|
|
return do_append(self, Py_SIZE(self->stack) - 1);
|
|
}
|
|
|
|
static int
|
|
load_appends(UnpicklerObject *self)
|
|
{
|
|
Py_ssize_t i = marker(self);
|
|
if (i < 0)
|
|
return -1;
|
|
return do_append(self, i);
|
|
}
|
|
|
|
static int
|
|
do_setitems(UnpicklerObject *self, Py_ssize_t x)
|
|
{
|
|
PyObject *value, *key;
|
|
PyObject *dict;
|
|
Py_ssize_t len, i;
|
|
int status = 0;
|
|
|
|
len = Py_SIZE(self->stack);
|
|
if (x > len || x <= self->stack->fence)
|
|
return Pdata_stack_underflow(self->stack);
|
|
if (len == x) /* nothing to do */
|
|
return 0;
|
|
if ((len - x) % 2 != 0) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
/* Currupt or hostile pickle -- we never write one like this. */
|
|
PyErr_SetString(st->UnpicklingError,
|
|
"odd number of items for SETITEMS");
|
|
return -1;
|
|
}
|
|
|
|
/* Here, dict does not actually need to be a PyDict; it could be anything
|
|
that supports the __setitem__ attribute. */
|
|
dict = self->stack->data[x - 1];
|
|
|
|
for (i = x + 1; i < len; i += 2) {
|
|
key = self->stack->data[i - 1];
|
|
value = self->stack->data[i];
|
|
if (PyObject_SetItem(dict, key, value) < 0) {
|
|
status = -1;
|
|
break;
|
|
}
|
|
}
|
|
|
|
Pdata_clear(self->stack, x);
|
|
return status;
|
|
}
|
|
|
|
static int
|
|
load_setitem(UnpicklerObject *self)
|
|
{
|
|
return do_setitems(self, Py_SIZE(self->stack) - 2);
|
|
}
|
|
|
|
static int
|
|
load_setitems(UnpicklerObject *self)
|
|
{
|
|
Py_ssize_t i = marker(self);
|
|
if (i < 0)
|
|
return -1;
|
|
return do_setitems(self, i);
|
|
}
|
|
|
|
static int
|
|
load_additems(UnpicklerObject *self)
|
|
{
|
|
PyObject *set;
|
|
Py_ssize_t mark, len, i;
|
|
|
|
mark = marker(self);
|
|
if (mark < 0)
|
|
return -1;
|
|
len = Py_SIZE(self->stack);
|
|
if (mark > len || mark <= self->stack->fence)
|
|
return Pdata_stack_underflow(self->stack);
|
|
if (len == mark) /* nothing to do */
|
|
return 0;
|
|
|
|
set = self->stack->data[mark - 1];
|
|
|
|
if (PySet_Check(set)) {
|
|
PyObject *items;
|
|
int status;
|
|
|
|
items = Pdata_poptuple(self->stack, mark);
|
|
if (items == NULL)
|
|
return -1;
|
|
|
|
status = _PySet_Update(set, items);
|
|
Py_DECREF(items);
|
|
return status;
|
|
}
|
|
else {
|
|
PyObject *add_func;
|
|
_Py_IDENTIFIER(add);
|
|
|
|
add_func = _PyObject_GetAttrId(set, &PyId_add);
|
|
if (add_func == NULL)
|
|
return -1;
|
|
for (i = mark; i < len; i++) {
|
|
PyObject *result;
|
|
PyObject *item;
|
|
|
|
item = self->stack->data[i];
|
|
result = _Pickle_FastCall(add_func, item);
|
|
if (result == NULL) {
|
|
Pdata_clear(self->stack, i + 1);
|
|
Py_SET_SIZE(self->stack, mark);
|
|
return -1;
|
|
}
|
|
Py_DECREF(result);
|
|
}
|
|
Py_SET_SIZE(self->stack, mark);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_build(UnpicklerObject *self)
|
|
{
|
|
PyObject *state, *inst, *slotstate;
|
|
PyObject *setstate;
|
|
int status = 0;
|
|
_Py_IDENTIFIER(__setstate__);
|
|
|
|
/* Stack is ... instance, state. We want to leave instance at
|
|
* the stack top, possibly mutated via instance.__setstate__(state).
|
|
*/
|
|
if (Py_SIZE(self->stack) - 2 < self->stack->fence)
|
|
return Pdata_stack_underflow(self->stack);
|
|
|
|
PDATA_POP(self->stack, state);
|
|
if (state == NULL)
|
|
return -1;
|
|
|
|
inst = self->stack->data[Py_SIZE(self->stack) - 1];
|
|
|
|
if (_PyObject_LookupAttrId(inst, &PyId___setstate__, &setstate) < 0) {
|
|
Py_DECREF(state);
|
|
return -1;
|
|
}
|
|
if (setstate != NULL) {
|
|
PyObject *result;
|
|
|
|
/* The explicit __setstate__ is responsible for everything. */
|
|
result = _Pickle_FastCall(setstate, state);
|
|
Py_DECREF(setstate);
|
|
if (result == NULL)
|
|
return -1;
|
|
Py_DECREF(result);
|
|
return 0;
|
|
}
|
|
|
|
/* A default __setstate__. First see whether state embeds a
|
|
* slot state dict too (a proto 2 addition).
|
|
*/
|
|
if (PyTuple_Check(state) && PyTuple_GET_SIZE(state) == 2) {
|
|
PyObject *tmp = state;
|
|
|
|
state = PyTuple_GET_ITEM(tmp, 0);
|
|
slotstate = PyTuple_GET_ITEM(tmp, 1);
|
|
Py_INCREF(state);
|
|
Py_INCREF(slotstate);
|
|
Py_DECREF(tmp);
|
|
}
|
|
else
|
|
slotstate = NULL;
|
|
|
|
/* Set inst.__dict__ from the state dict (if any). */
|
|
if (state != Py_None) {
|
|
PyObject *dict;
|
|
PyObject *d_key, *d_value;
|
|
Py_ssize_t i;
|
|
_Py_IDENTIFIER(__dict__);
|
|
|
|
if (!PyDict_Check(state)) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->UnpicklingError, "state is not a dictionary");
|
|
goto error;
|
|
}
|
|
dict = _PyObject_GetAttrId(inst, &PyId___dict__);
|
|
if (dict == NULL)
|
|
goto error;
|
|
|
|
i = 0;
|
|
while (PyDict_Next(state, &i, &d_key, &d_value)) {
|
|
/* normally the keys for instance attributes are
|
|
interned. we should try to do that here. */
|
|
Py_INCREF(d_key);
|
|
if (PyUnicode_CheckExact(d_key))
|
|
PyUnicode_InternInPlace(&d_key);
|
|
if (PyObject_SetItem(dict, d_key, d_value) < 0) {
|
|
Py_DECREF(d_key);
|
|
goto error;
|
|
}
|
|
Py_DECREF(d_key);
|
|
}
|
|
Py_DECREF(dict);
|
|
}
|
|
|
|
/* Also set instance attributes from the slotstate dict (if any). */
|
|
if (slotstate != NULL) {
|
|
PyObject *d_key, *d_value;
|
|
Py_ssize_t i;
|
|
|
|
if (!PyDict_Check(slotstate)) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_SetString(st->UnpicklingError,
|
|
"slot state is not a dictionary");
|
|
goto error;
|
|
}
|
|
i = 0;
|
|
while (PyDict_Next(slotstate, &i, &d_key, &d_value)) {
|
|
if (PyObject_SetAttr(inst, d_key, d_value) < 0)
|
|
goto error;
|
|
}
|
|
}
|
|
|
|
if (0) {
|
|
error:
|
|
status = -1;
|
|
}
|
|
|
|
Py_DECREF(state);
|
|
Py_XDECREF(slotstate);
|
|
return status;
|
|
}
|
|
|
|
static int
|
|
load_mark(UnpicklerObject *self)
|
|
{
|
|
|
|
/* Note that we split the (pickle.py) stack into two stacks, an
|
|
* object stack and a mark stack. Here we push a mark onto the
|
|
* mark stack.
|
|
*/
|
|
|
|
if (self->num_marks >= self->marks_size) {
|
|
size_t alloc = ((size_t)self->num_marks << 1) + 20;
|
|
Py_ssize_t *marks_new = self->marks;
|
|
PyMem_RESIZE(marks_new, Py_ssize_t, alloc);
|
|
if (marks_new == NULL) {
|
|
PyErr_NoMemory();
|
|
return -1;
|
|
}
|
|
self->marks = marks_new;
|
|
self->marks_size = (Py_ssize_t)alloc;
|
|
}
|
|
|
|
self->stack->mark_set = 1;
|
|
self->marks[self->num_marks++] = self->stack->fence = Py_SIZE(self->stack);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
load_reduce(UnpicklerObject *self)
|
|
{
|
|
PyObject *callable = NULL;
|
|
PyObject *argtup = NULL;
|
|
PyObject *obj = NULL;
|
|
|
|
PDATA_POP(self->stack, argtup);
|
|
if (argtup == NULL)
|
|
return -1;
|
|
PDATA_POP(self->stack, callable);
|
|
if (callable) {
|
|
obj = PyObject_CallObject(callable, argtup);
|
|
Py_DECREF(callable);
|
|
}
|
|
Py_DECREF(argtup);
|
|
|
|
if (obj == NULL)
|
|
return -1;
|
|
|
|
PDATA_PUSH(self->stack, obj, -1);
|
|
return 0;
|
|
}
|
|
|
|
/* Just raises an error if we don't know the protocol specified. PROTO
|
|
* is the first opcode for protocols >= 2.
|
|
*/
|
|
static int
|
|
load_proto(UnpicklerObject *self)
|
|
{
|
|
char *s;
|
|
int i;
|
|
|
|
if (_Unpickler_Read(self, &s, 1) < 0)
|
|
return -1;
|
|
|
|
i = (unsigned char)s[0];
|
|
if (i <= HIGHEST_PROTOCOL) {
|
|
self->proto = i;
|
|
return 0;
|
|
}
|
|
|
|
PyErr_Format(PyExc_ValueError, "unsupported pickle protocol: %d", i);
|
|
return -1;
|
|
}
|
|
|
|
static int
|
|
load_frame(UnpicklerObject *self)
|
|
{
|
|
char *s;
|
|
Py_ssize_t frame_len;
|
|
|
|
if (_Unpickler_Read(self, &s, 8) < 0)
|
|
return -1;
|
|
|
|
frame_len = calc_binsize(s, 8);
|
|
if (frame_len < 0) {
|
|
PyErr_Format(PyExc_OverflowError,
|
|
"FRAME length exceeds system's maximum of %zd bytes",
|
|
PY_SSIZE_T_MAX);
|
|
return -1;
|
|
}
|
|
|
|
if (_Unpickler_Read(self, &s, frame_len) < 0)
|
|
return -1;
|
|
|
|
/* Rewind to start of frame */
|
|
self->next_read_idx -= frame_len;
|
|
return 0;
|
|
}
|
|
|
|
static PyObject *
|
|
load(UnpicklerObject *self)
|
|
{
|
|
PyObject *value = NULL;
|
|
char *s = NULL;
|
|
|
|
self->num_marks = 0;
|
|
self->stack->mark_set = 0;
|
|
self->stack->fence = 0;
|
|
self->proto = 0;
|
|
if (Py_SIZE(self->stack))
|
|
Pdata_clear(self->stack, 0);
|
|
|
|
/* Convenient macros for the dispatch while-switch loop just below. */
|
|
#define OP(opcode, load_func) \
|
|
case opcode: if (load_func(self) < 0) break; continue;
|
|
|
|
#define OP_ARG(opcode, load_func, arg) \
|
|
case opcode: if (load_func(self, (arg)) < 0) break; continue;
|
|
|
|
while (1) {
|
|
if (_Unpickler_Read(self, &s, 1) < 0) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
if (PyErr_ExceptionMatches(st->UnpicklingError)) {
|
|
PyErr_Format(PyExc_EOFError, "Ran out of input");
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
switch ((enum opcode)s[0]) {
|
|
OP(NONE, load_none)
|
|
OP(BININT, load_binint)
|
|
OP(BININT1, load_binint1)
|
|
OP(BININT2, load_binint2)
|
|
OP(INT, load_int)
|
|
OP(LONG, load_long)
|
|
OP_ARG(LONG1, load_counted_long, 1)
|
|
OP_ARG(LONG4, load_counted_long, 4)
|
|
OP(FLOAT, load_float)
|
|
OP(BINFLOAT, load_binfloat)
|
|
OP_ARG(SHORT_BINBYTES, load_counted_binbytes, 1)
|
|
OP_ARG(BINBYTES, load_counted_binbytes, 4)
|
|
OP_ARG(BINBYTES8, load_counted_binbytes, 8)
|
|
OP(BYTEARRAY8, load_counted_bytearray)
|
|
OP(NEXT_BUFFER, load_next_buffer)
|
|
OP(READONLY_BUFFER, load_readonly_buffer)
|
|
OP_ARG(SHORT_BINSTRING, load_counted_binstring, 1)
|
|
OP_ARG(BINSTRING, load_counted_binstring, 4)
|
|
OP(STRING, load_string)
|
|
OP(UNICODE, load_unicode)
|
|
OP_ARG(SHORT_BINUNICODE, load_counted_binunicode, 1)
|
|
OP_ARG(BINUNICODE, load_counted_binunicode, 4)
|
|
OP_ARG(BINUNICODE8, load_counted_binunicode, 8)
|
|
OP_ARG(EMPTY_TUPLE, load_counted_tuple, 0)
|
|
OP_ARG(TUPLE1, load_counted_tuple, 1)
|
|
OP_ARG(TUPLE2, load_counted_tuple, 2)
|
|
OP_ARG(TUPLE3, load_counted_tuple, 3)
|
|
OP(TUPLE, load_tuple)
|
|
OP(EMPTY_LIST, load_empty_list)
|
|
OP(LIST, load_list)
|
|
OP(EMPTY_DICT, load_empty_dict)
|
|
OP(DICT, load_dict)
|
|
OP(EMPTY_SET, load_empty_set)
|
|
OP(ADDITEMS, load_additems)
|
|
OP(FROZENSET, load_frozenset)
|
|
OP(OBJ, load_obj)
|
|
OP(INST, load_inst)
|
|
OP_ARG(NEWOBJ, load_newobj, 0)
|
|
OP_ARG(NEWOBJ_EX, load_newobj, 1)
|
|
OP(GLOBAL, load_global)
|
|
OP(STACK_GLOBAL, load_stack_global)
|
|
OP(APPEND, load_append)
|
|
OP(APPENDS, load_appends)
|
|
OP(BUILD, load_build)
|
|
OP(DUP, load_dup)
|
|
OP(BINGET, load_binget)
|
|
OP(LONG_BINGET, load_long_binget)
|
|
OP(GET, load_get)
|
|
OP(MARK, load_mark)
|
|
OP(BINPUT, load_binput)
|
|
OP(LONG_BINPUT, load_long_binput)
|
|
OP(PUT, load_put)
|
|
OP(MEMOIZE, load_memoize)
|
|
OP(POP, load_pop)
|
|
OP(POP_MARK, load_pop_mark)
|
|
OP(SETITEM, load_setitem)
|
|
OP(SETITEMS, load_setitems)
|
|
OP(PERSID, load_persid)
|
|
OP(BINPERSID, load_binpersid)
|
|
OP(REDUCE, load_reduce)
|
|
OP(PROTO, load_proto)
|
|
OP(FRAME, load_frame)
|
|
OP_ARG(EXT1, load_extension, 1)
|
|
OP_ARG(EXT2, load_extension, 2)
|
|
OP_ARG(EXT4, load_extension, 4)
|
|
OP_ARG(NEWTRUE, load_bool, Py_True)
|
|
OP_ARG(NEWFALSE, load_bool, Py_False)
|
|
|
|
case STOP:
|
|
break;
|
|
|
|
default:
|
|
{
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
unsigned char c = (unsigned char) *s;
|
|
if (0x20 <= c && c <= 0x7e && c != '\'' && c != '\\') {
|
|
PyErr_Format(st->UnpicklingError,
|
|
"invalid load key, '%c'.", c);
|
|
}
|
|
else {
|
|
PyErr_Format(st->UnpicklingError,
|
|
"invalid load key, '\\x%02x'.", c);
|
|
}
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
break; /* and we are done! */
|
|
}
|
|
|
|
if (PyErr_Occurred()) {
|
|
return NULL;
|
|
}
|
|
|
|
if (_Unpickler_SkipConsumed(self) < 0)
|
|
return NULL;
|
|
|
|
PDATA_POP(self->stack, value);
|
|
return value;
|
|
}
|
|
|
|
/*[clinic input]
|
|
|
|
_pickle.Unpickler.load
|
|
|
|
Load a pickle.
|
|
|
|
Read a pickled object representation from the open file object given
|
|
in the constructor, and return the reconstituted object hierarchy
|
|
specified therein.
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_Unpickler_load_impl(UnpicklerObject *self)
|
|
/*[clinic end generated code: output=fdcc488aad675b14 input=acbb91a42fa9b7b9]*/
|
|
{
|
|
UnpicklerObject *unpickler = (UnpicklerObject*)self;
|
|
|
|
/* Check whether the Unpickler was initialized correctly. This prevents
|
|
segfaulting if a subclass overridden __init__ with a function that does
|
|
not call Unpickler.__init__(). Here, we simply ensure that self->read
|
|
is not NULL. */
|
|
if (unpickler->read == NULL) {
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
PyErr_Format(st->UnpicklingError,
|
|
"Unpickler.__init__() was not called by %s.__init__()",
|
|
Py_TYPE(unpickler)->tp_name);
|
|
return NULL;
|
|
}
|
|
|
|
return load(unpickler);
|
|
}
|
|
|
|
/* The name of find_class() is misleading. In newer pickle protocols, this
|
|
function is used for loading any global (i.e., functions), not just
|
|
classes. The name is kept only for backward compatibility. */
|
|
|
|
/*[clinic input]
|
|
|
|
_pickle.Unpickler.find_class
|
|
|
|
module_name: object
|
|
global_name: object
|
|
/
|
|
|
|
Return an object from a specified module.
|
|
|
|
If necessary, the module will be imported. Subclasses may override
|
|
this method (e.g. to restrict unpickling of arbitrary classes and
|
|
functions).
|
|
|
|
This method is called whenever a class or a function object is
|
|
needed. Both arguments passed are str objects.
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_Unpickler_find_class_impl(UnpicklerObject *self,
|
|
PyObject *module_name,
|
|
PyObject *global_name)
|
|
/*[clinic end generated code: output=becc08d7f9ed41e3 input=e2e6a865de093ef4]*/
|
|
{
|
|
PyObject *global;
|
|
PyObject *module;
|
|
|
|
if (PySys_Audit("pickle.find_class", "OO",
|
|
module_name, global_name) < 0) {
|
|
return NULL;
|
|
}
|
|
|
|
/* Try to map the old names used in Python 2.x to the new ones used in
|
|
Python 3.x. We do this only with old pickle protocols and when the
|
|
user has not disabled the feature. */
|
|
if (self->proto < 3 && self->fix_imports) {
|
|
PyObject *key;
|
|
PyObject *item;
|
|
PickleState *st = _Pickle_GetGlobalState();
|
|
|
|
/* Check if the global (i.e., a function or a class) was renamed
|
|
or moved to another module. */
|
|
key = PyTuple_Pack(2, module_name, global_name);
|
|
if (key == NULL)
|
|
return NULL;
|
|
item = PyDict_GetItemWithError(st->name_mapping_2to3, key);
|
|
Py_DECREF(key);
|
|
if (item) {
|
|
if (!PyTuple_Check(item) || PyTuple_GET_SIZE(item) != 2) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"_compat_pickle.NAME_MAPPING values should be "
|
|
"2-tuples, not %.200s", Py_TYPE(item)->tp_name);
|
|
return NULL;
|
|
}
|
|
module_name = PyTuple_GET_ITEM(item, 0);
|
|
global_name = PyTuple_GET_ITEM(item, 1);
|
|
if (!PyUnicode_Check(module_name) ||
|
|
!PyUnicode_Check(global_name)) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"_compat_pickle.NAME_MAPPING values should be "
|
|
"pairs of str, not (%.200s, %.200s)",
|
|
Py_TYPE(module_name)->tp_name,
|
|
Py_TYPE(global_name)->tp_name);
|
|
return NULL;
|
|
}
|
|
}
|
|
else if (PyErr_Occurred()) {
|
|
return NULL;
|
|
}
|
|
else {
|
|
/* Check if the module was renamed. */
|
|
item = PyDict_GetItemWithError(st->import_mapping_2to3, module_name);
|
|
if (item) {
|
|
if (!PyUnicode_Check(item)) {
|
|
PyErr_Format(PyExc_RuntimeError,
|
|
"_compat_pickle.IMPORT_MAPPING values should be "
|
|
"strings, not %.200s", Py_TYPE(item)->tp_name);
|
|
return NULL;
|
|
}
|
|
module_name = item;
|
|
}
|
|
else if (PyErr_Occurred()) {
|
|
return NULL;
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* we don't use PyImport_GetModule here, because it can return partially-
|
|
* initialised modules, which then cause the getattribute to fail.
|
|
*/
|
|
module = PyImport_Import(module_name);
|
|
if (module == NULL) {
|
|
return NULL;
|
|
}
|
|
global = getattribute(module, global_name, self->proto >= 4);
|
|
Py_DECREF(module);
|
|
return global;
|
|
}
|
|
|
|
/*[clinic input]
|
|
|
|
_pickle.Unpickler.__sizeof__ -> Py_ssize_t
|
|
|
|
Returns size in memory, in bytes.
|
|
[clinic start generated code]*/
|
|
|
|
static Py_ssize_t
|
|
_pickle_Unpickler___sizeof___impl(UnpicklerObject *self)
|
|
/*[clinic end generated code: output=119d9d03ad4c7651 input=13333471fdeedf5e]*/
|
|
{
|
|
Py_ssize_t res;
|
|
|
|
res = _PyObject_SIZE(Py_TYPE(self));
|
|
if (self->memo != NULL)
|
|
res += self->memo_size * sizeof(PyObject *);
|
|
if (self->marks != NULL)
|
|
res += self->marks_size * sizeof(Py_ssize_t);
|
|
if (self->input_line != NULL)
|
|
res += strlen(self->input_line) + 1;
|
|
if (self->encoding != NULL)
|
|
res += strlen(self->encoding) + 1;
|
|
if (self->errors != NULL)
|
|
res += strlen(self->errors) + 1;
|
|
return res;
|
|
}
|
|
|
|
static struct PyMethodDef Unpickler_methods[] = {
|
|
_PICKLE_UNPICKLER_LOAD_METHODDEF
|
|
_PICKLE_UNPICKLER_FIND_CLASS_METHODDEF
|
|
_PICKLE_UNPICKLER___SIZEOF___METHODDEF
|
|
{NULL, NULL} /* sentinel */
|
|
};
|
|
|
|
static void
|
|
Unpickler_dealloc(UnpicklerObject *self)
|
|
{
|
|
PyObject_GC_UnTrack((PyObject *)self);
|
|
Py_XDECREF(self->readline);
|
|
Py_XDECREF(self->readinto);
|
|
Py_XDECREF(self->read);
|
|
Py_XDECREF(self->peek);
|
|
Py_XDECREF(self->stack);
|
|
Py_XDECREF(self->pers_func);
|
|
Py_XDECREF(self->buffers);
|
|
if (self->buffer.buf != NULL) {
|
|
PyBuffer_Release(&self->buffer);
|
|
self->buffer.buf = NULL;
|
|
}
|
|
|
|
_Unpickler_MemoCleanup(self);
|
|
PyMem_Free(self->marks);
|
|
PyMem_Free(self->input_line);
|
|
PyMem_Free(self->encoding);
|
|
PyMem_Free(self->errors);
|
|
|
|
Py_TYPE(self)->tp_free((PyObject *)self);
|
|
}
|
|
|
|
static int
|
|
Unpickler_traverse(UnpicklerObject *self, visitproc visit, void *arg)
|
|
{
|
|
Py_VISIT(self->readline);
|
|
Py_VISIT(self->readinto);
|
|
Py_VISIT(self->read);
|
|
Py_VISIT(self->peek);
|
|
Py_VISIT(self->stack);
|
|
Py_VISIT(self->pers_func);
|
|
Py_VISIT(self->buffers);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
Unpickler_clear(UnpicklerObject *self)
|
|
{
|
|
Py_CLEAR(self->readline);
|
|
Py_CLEAR(self->readinto);
|
|
Py_CLEAR(self->read);
|
|
Py_CLEAR(self->peek);
|
|
Py_CLEAR(self->stack);
|
|
Py_CLEAR(self->pers_func);
|
|
Py_CLEAR(self->buffers);
|
|
if (self->buffer.buf != NULL) {
|
|
PyBuffer_Release(&self->buffer);
|
|
self->buffer.buf = NULL;
|
|
}
|
|
|
|
_Unpickler_MemoCleanup(self);
|
|
PyMem_Free(self->marks);
|
|
self->marks = NULL;
|
|
PyMem_Free(self->input_line);
|
|
self->input_line = NULL;
|
|
PyMem_Free(self->encoding);
|
|
self->encoding = NULL;
|
|
PyMem_Free(self->errors);
|
|
self->errors = NULL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*[clinic input]
|
|
|
|
_pickle.Unpickler.__init__
|
|
|
|
file: object
|
|
*
|
|
fix_imports: bool = True
|
|
encoding: str = 'ASCII'
|
|
errors: str = 'strict'
|
|
buffers: object(c_default="NULL") = ()
|
|
|
|
This takes a binary file for reading a pickle data stream.
|
|
|
|
The protocol version of the pickle is detected automatically, so no
|
|
protocol argument is needed. Bytes past the pickled object's
|
|
representation are ignored.
|
|
|
|
The argument *file* must have two methods, a read() method that takes
|
|
an integer argument, and a readline() method that requires no
|
|
arguments. Both methods should return bytes. Thus *file* can be a
|
|
binary file object opened for reading, an io.BytesIO object, or any
|
|
other custom object that meets this interface.
|
|
|
|
Optional keyword arguments are *fix_imports*, *encoding* and *errors*,
|
|
which are used to control compatibility support for pickle stream
|
|
generated by Python 2. If *fix_imports* is True, pickle will try to
|
|
map the old Python 2 names to the new names used in Python 3. The
|
|
*encoding* and *errors* tell pickle how to decode 8-bit string
|
|
instances pickled by Python 2; these default to 'ASCII' and 'strict',
|
|
respectively. The *encoding* can be 'bytes' to read these 8-bit
|
|
string instances as bytes objects.
|
|
[clinic start generated code]*/
|
|
|
|
static int
|
|
_pickle_Unpickler___init___impl(UnpicklerObject *self, PyObject *file,
|
|
int fix_imports, const char *encoding,
|
|
const char *errors, PyObject *buffers)
|
|
/*[clinic end generated code: output=09f0192649ea3f85 input=ca4c1faea9553121]*/
|
|
{
|
|
_Py_IDENTIFIER(persistent_load);
|
|
|
|
/* In case of multiple __init__() calls, clear previous content. */
|
|
if (self->read != NULL)
|
|
(void)Unpickler_clear(self);
|
|
|
|
if (_Unpickler_SetInputStream(self, file) < 0)
|
|
return -1;
|
|
|
|
if (_Unpickler_SetInputEncoding(self, encoding, errors) < 0)
|
|
return -1;
|
|
|
|
if (_Unpickler_SetBuffers(self, buffers) < 0)
|
|
return -1;
|
|
|
|
self->fix_imports = fix_imports;
|
|
|
|
if (init_method_ref((PyObject *)self, &PyId_persistent_load,
|
|
&self->pers_func, &self->pers_func_self) < 0)
|
|
{
|
|
return -1;
|
|
}
|
|
|
|
self->stack = (Pdata *)Pdata_New();
|
|
if (self->stack == NULL)
|
|
return -1;
|
|
|
|
self->memo_size = 32;
|
|
self->memo = _Unpickler_NewMemo(self->memo_size);
|
|
if (self->memo == NULL)
|
|
return -1;
|
|
|
|
self->proto = 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
/* Define a proxy object for the Unpickler's internal memo object. This is to
|
|
* avoid breaking code like:
|
|
* unpickler.memo.clear()
|
|
* and
|
|
* unpickler.memo = saved_memo
|
|
* Is this a good idea? Not really, but we don't want to break code that uses
|
|
* it. Note that we don't implement the entire mapping API here. This is
|
|
* intentional, as these should be treated as black-box implementation details.
|
|
*
|
|
* We do, however, have to implement pickling/unpickling support because of
|
|
* real-world code like cvs2svn.
|
|
*/
|
|
|
|
/*[clinic input]
|
|
_pickle.UnpicklerMemoProxy.clear
|
|
|
|
Remove all items from memo.
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_UnpicklerMemoProxy_clear_impl(UnpicklerMemoProxyObject *self)
|
|
/*[clinic end generated code: output=d20cd43f4ba1fb1f input=b1df7c52e7afd9bd]*/
|
|
{
|
|
_Unpickler_MemoCleanup(self->unpickler);
|
|
self->unpickler->memo = _Unpickler_NewMemo(self->unpickler->memo_size);
|
|
if (self->unpickler->memo == NULL)
|
|
return NULL;
|
|
Py_RETURN_NONE;
|
|
}
|
|
|
|
/*[clinic input]
|
|
_pickle.UnpicklerMemoProxy.copy
|
|
|
|
Copy the memo to a new object.
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_UnpicklerMemoProxy_copy_impl(UnpicklerMemoProxyObject *self)
|
|
/*[clinic end generated code: output=e12af7e9bc1e4c77 input=97769247ce032c1d]*/
|
|
{
|
|
size_t i;
|
|
PyObject *new_memo = PyDict_New();
|
|
if (new_memo == NULL)
|
|
return NULL;
|
|
|
|
for (i = 0; i < self->unpickler->memo_size; i++) {
|
|
int status;
|
|
PyObject *key, *value;
|
|
|
|
value = self->unpickler->memo[i];
|
|
if (value == NULL)
|
|
continue;
|
|
|
|
key = PyLong_FromSsize_t(i);
|
|
if (key == NULL)
|
|
goto error;
|
|
status = PyDict_SetItem(new_memo, key, value);
|
|
Py_DECREF(key);
|
|
if (status < 0)
|
|
goto error;
|
|
}
|
|
return new_memo;
|
|
|
|
error:
|
|
Py_DECREF(new_memo);
|
|
return NULL;
|
|
}
|
|
|
|
/*[clinic input]
|
|
_pickle.UnpicklerMemoProxy.__reduce__
|
|
|
|
Implement pickling support.
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_UnpicklerMemoProxy___reduce___impl(UnpicklerMemoProxyObject *self)
|
|
/*[clinic end generated code: output=6da34ac048d94cca input=6920862413407199]*/
|
|
{
|
|
PyObject *reduce_value;
|
|
PyObject *constructor_args;
|
|
PyObject *contents = _pickle_UnpicklerMemoProxy_copy_impl(self);
|
|
if (contents == NULL)
|
|
return NULL;
|
|
|
|
reduce_value = PyTuple_New(2);
|
|
if (reduce_value == NULL) {
|
|
Py_DECREF(contents);
|
|
return NULL;
|
|
}
|
|
constructor_args = PyTuple_New(1);
|
|
if (constructor_args == NULL) {
|
|
Py_DECREF(contents);
|
|
Py_DECREF(reduce_value);
|
|
return NULL;
|
|
}
|
|
PyTuple_SET_ITEM(constructor_args, 0, contents);
|
|
Py_INCREF((PyObject *)&PyDict_Type);
|
|
PyTuple_SET_ITEM(reduce_value, 0, (PyObject *)&PyDict_Type);
|
|
PyTuple_SET_ITEM(reduce_value, 1, constructor_args);
|
|
return reduce_value;
|
|
}
|
|
|
|
static PyMethodDef unpicklerproxy_methods[] = {
|
|
_PICKLE_UNPICKLERMEMOPROXY_CLEAR_METHODDEF
|
|
_PICKLE_UNPICKLERMEMOPROXY_COPY_METHODDEF
|
|
_PICKLE_UNPICKLERMEMOPROXY___REDUCE___METHODDEF
|
|
{NULL, NULL} /* sentinel */
|
|
};
|
|
|
|
static void
|
|
UnpicklerMemoProxy_dealloc(UnpicklerMemoProxyObject *self)
|
|
{
|
|
PyObject_GC_UnTrack(self);
|
|
Py_XDECREF(self->unpickler);
|
|
PyObject_GC_Del((PyObject *)self);
|
|
}
|
|
|
|
static int
|
|
UnpicklerMemoProxy_traverse(UnpicklerMemoProxyObject *self,
|
|
visitproc visit, void *arg)
|
|
{
|
|
Py_VISIT(self->unpickler);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
UnpicklerMemoProxy_clear(UnpicklerMemoProxyObject *self)
|
|
{
|
|
Py_CLEAR(self->unpickler);
|
|
return 0;
|
|
}
|
|
|
|
static PyTypeObject UnpicklerMemoProxyType = {
|
|
PyVarObject_HEAD_INIT(NULL, 0)
|
|
"_pickle.UnpicklerMemoProxy", /*tp_name*/
|
|
sizeof(UnpicklerMemoProxyObject), /*tp_basicsize*/
|
|
0,
|
|
(destructor)UnpicklerMemoProxy_dealloc, /* tp_dealloc */
|
|
0, /* tp_vectorcall_offset */
|
|
0, /* tp_getattr */
|
|
0, /* tp_setattr */
|
|
0, /* tp_as_async */
|
|
0, /* tp_repr */
|
|
0, /* tp_as_number */
|
|
0, /* tp_as_sequence */
|
|
0, /* tp_as_mapping */
|
|
PyObject_HashNotImplemented, /* tp_hash */
|
|
0, /* tp_call */
|
|
0, /* tp_str */
|
|
PyObject_GenericGetAttr, /* tp_getattro */
|
|
PyObject_GenericSetAttr, /* tp_setattro */
|
|
0, /* tp_as_buffer */
|
|
Py_TPFLAGS_DEFAULT | Py_TPFLAGS_BASETYPE | Py_TPFLAGS_HAVE_GC,
|
|
0, /* tp_doc */
|
|
(traverseproc)UnpicklerMemoProxy_traverse, /* tp_traverse */
|
|
(inquiry)UnpicklerMemoProxy_clear, /* tp_clear */
|
|
0, /* tp_richcompare */
|
|
0, /* tp_weaklistoffset */
|
|
0, /* tp_iter */
|
|
0, /* tp_iternext */
|
|
unpicklerproxy_methods, /* tp_methods */
|
|
};
|
|
|
|
static PyObject *
|
|
UnpicklerMemoProxy_New(UnpicklerObject *unpickler)
|
|
{
|
|
UnpicklerMemoProxyObject *self;
|
|
|
|
self = PyObject_GC_New(UnpicklerMemoProxyObject,
|
|
&UnpicklerMemoProxyType);
|
|
if (self == NULL)
|
|
return NULL;
|
|
Py_INCREF(unpickler);
|
|
self->unpickler = unpickler;
|
|
PyObject_GC_Track(self);
|
|
return (PyObject *)self;
|
|
}
|
|
|
|
/*****************************************************************************/
|
|
|
|
|
|
static PyObject *
|
|
Unpickler_get_memo(UnpicklerObject *self, void *Py_UNUSED(ignored))
|
|
{
|
|
return UnpicklerMemoProxy_New(self);
|
|
}
|
|
|
|
static int
|
|
Unpickler_set_memo(UnpicklerObject *self, PyObject *obj, void *Py_UNUSED(ignored))
|
|
{
|
|
PyObject **new_memo;
|
|
size_t new_memo_size = 0;
|
|
|
|
if (obj == NULL) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"attribute deletion is not supported");
|
|
return -1;
|
|
}
|
|
|
|
if (Py_IS_TYPE(obj, &UnpicklerMemoProxyType)) {
|
|
UnpicklerObject *unpickler =
|
|
((UnpicklerMemoProxyObject *)obj)->unpickler;
|
|
|
|
new_memo_size = unpickler->memo_size;
|
|
new_memo = _Unpickler_NewMemo(new_memo_size);
|
|
if (new_memo == NULL)
|
|
return -1;
|
|
|
|
for (size_t i = 0; i < new_memo_size; i++) {
|
|
Py_XINCREF(unpickler->memo[i]);
|
|
new_memo[i] = unpickler->memo[i];
|
|
}
|
|
}
|
|
else if (PyDict_Check(obj)) {
|
|
Py_ssize_t i = 0;
|
|
PyObject *key, *value;
|
|
|
|
new_memo_size = PyDict_GET_SIZE(obj);
|
|
new_memo = _Unpickler_NewMemo(new_memo_size);
|
|
if (new_memo == NULL)
|
|
return -1;
|
|
|
|
while (PyDict_Next(obj, &i, &key, &value)) {
|
|
Py_ssize_t idx;
|
|
if (!PyLong_Check(key)) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"memo key must be integers");
|
|
goto error;
|
|
}
|
|
idx = PyLong_AsSsize_t(key);
|
|
if (idx == -1 && PyErr_Occurred())
|
|
goto error;
|
|
if (idx < 0) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"memo key must be positive integers.");
|
|
goto error;
|
|
}
|
|
if (_Unpickler_MemoPut(self, idx, value) < 0)
|
|
goto error;
|
|
}
|
|
}
|
|
else {
|
|
PyErr_Format(PyExc_TypeError,
|
|
"'memo' attribute must be an UnpicklerMemoProxy object "
|
|
"or dict, not %.200s", Py_TYPE(obj)->tp_name);
|
|
return -1;
|
|
}
|
|
|
|
_Unpickler_MemoCleanup(self);
|
|
self->memo_size = new_memo_size;
|
|
self->memo = new_memo;
|
|
|
|
return 0;
|
|
|
|
error:
|
|
if (new_memo_size) {
|
|
for (size_t i = new_memo_size - 1; i != SIZE_MAX; i--) {
|
|
Py_XDECREF(new_memo[i]);
|
|
}
|
|
PyMem_FREE(new_memo);
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
static PyObject *
|
|
Unpickler_get_persload(UnpicklerObject *self, void *Py_UNUSED(ignored))
|
|
{
|
|
if (self->pers_func == NULL) {
|
|
PyErr_SetString(PyExc_AttributeError, "persistent_load");
|
|
return NULL;
|
|
}
|
|
return reconstruct_method(self->pers_func, self->pers_func_self);
|
|
}
|
|
|
|
static int
|
|
Unpickler_set_persload(UnpicklerObject *self, PyObject *value, void *Py_UNUSED(ignored))
|
|
{
|
|
if (value == NULL) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"attribute deletion is not supported");
|
|
return -1;
|
|
}
|
|
if (!PyCallable_Check(value)) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"persistent_load must be a callable taking "
|
|
"one argument");
|
|
return -1;
|
|
}
|
|
|
|
self->pers_func_self = NULL;
|
|
Py_INCREF(value);
|
|
Py_XSETREF(self->pers_func, value);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static PyGetSetDef Unpickler_getsets[] = {
|
|
{"memo", (getter)Unpickler_get_memo, (setter)Unpickler_set_memo},
|
|
{"persistent_load", (getter)Unpickler_get_persload,
|
|
(setter)Unpickler_set_persload},
|
|
{NULL}
|
|
};
|
|
|
|
static PyTypeObject Unpickler_Type = {
|
|
PyVarObject_HEAD_INIT(NULL, 0)
|
|
"_pickle.Unpickler", /*tp_name*/
|
|
sizeof(UnpicklerObject), /*tp_basicsize*/
|
|
0, /*tp_itemsize*/
|
|
(destructor)Unpickler_dealloc, /*tp_dealloc*/
|
|
0, /*tp_vectorcall_offset*/
|
|
0, /*tp_getattr*/
|
|
0, /*tp_setattr*/
|
|
0, /*tp_as_async*/
|
|
0, /*tp_repr*/
|
|
0, /*tp_as_number*/
|
|
0, /*tp_as_sequence*/
|
|
0, /*tp_as_mapping*/
|
|
0, /*tp_hash*/
|
|
0, /*tp_call*/
|
|
0, /*tp_str*/
|
|
0, /*tp_getattro*/
|
|
0, /*tp_setattro*/
|
|
0, /*tp_as_buffer*/
|
|
Py_TPFLAGS_DEFAULT | Py_TPFLAGS_BASETYPE | Py_TPFLAGS_HAVE_GC,
|
|
_pickle_Unpickler___init____doc__, /*tp_doc*/
|
|
(traverseproc)Unpickler_traverse, /*tp_traverse*/
|
|
(inquiry)Unpickler_clear, /*tp_clear*/
|
|
0, /*tp_richcompare*/
|
|
0, /*tp_weaklistoffset*/
|
|
0, /*tp_iter*/
|
|
0, /*tp_iternext*/
|
|
Unpickler_methods, /*tp_methods*/
|
|
0, /*tp_members*/
|
|
Unpickler_getsets, /*tp_getset*/
|
|
0, /*tp_base*/
|
|
0, /*tp_dict*/
|
|
0, /*tp_descr_get*/
|
|
0, /*tp_descr_set*/
|
|
0, /*tp_dictoffset*/
|
|
_pickle_Unpickler___init__, /*tp_init*/
|
|
PyType_GenericAlloc, /*tp_alloc*/
|
|
PyType_GenericNew, /*tp_new*/
|
|
PyObject_GC_Del, /*tp_free*/
|
|
0, /*tp_is_gc*/
|
|
};
|
|
|
|
/*[clinic input]
|
|
|
|
_pickle.dump
|
|
|
|
obj: object
|
|
file: object
|
|
protocol: object = None
|
|
*
|
|
fix_imports: bool = True
|
|
buffer_callback: object = None
|
|
|
|
Write a pickled representation of obj to the open file object file.
|
|
|
|
This is equivalent to ``Pickler(file, protocol).dump(obj)``, but may
|
|
be more efficient.
|
|
|
|
The optional *protocol* argument tells the pickler to use the given
|
|
protocol; supported protocols are 0, 1, 2, 3, 4 and 5. The default
|
|
protocol is 4. It was introduced in Python 3.4, and is incompatible
|
|
with previous versions.
|
|
|
|
Specifying a negative protocol version selects the highest protocol
|
|
version supported. The higher the protocol used, the more recent the
|
|
version of Python needed to read the pickle produced.
|
|
|
|
The *file* argument must have a write() method that accepts a single
|
|
bytes argument. It can thus be a file object opened for binary
|
|
writing, an io.BytesIO instance, or any other custom object that meets
|
|
this interface.
|
|
|
|
If *fix_imports* is True and protocol is less than 3, pickle will try
|
|
to map the new Python 3 names to the old module names used in Python
|
|
2, so that the pickle data stream is readable with Python 2.
|
|
|
|
If *buffer_callback* is None (the default), buffer views are serialized
|
|
into *file* as part of the pickle stream. It is an error if
|
|
*buffer_callback* is not None and *protocol* is None or smaller than 5.
|
|
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_dump_impl(PyObject *module, PyObject *obj, PyObject *file,
|
|
PyObject *protocol, int fix_imports,
|
|
PyObject *buffer_callback)
|
|
/*[clinic end generated code: output=706186dba996490c input=5ed6653da99cd97c]*/
|
|
{
|
|
PicklerObject *pickler = _Pickler_New();
|
|
|
|
if (pickler == NULL)
|
|
return NULL;
|
|
|
|
if (_Pickler_SetProtocol(pickler, protocol, fix_imports) < 0)
|
|
goto error;
|
|
|
|
if (_Pickler_SetOutputStream(pickler, file) < 0)
|
|
goto error;
|
|
|
|
if (_Pickler_SetBufferCallback(pickler, buffer_callback) < 0)
|
|
goto error;
|
|
|
|
if (dump(pickler, obj) < 0)
|
|
goto error;
|
|
|
|
if (_Pickler_FlushToFile(pickler) < 0)
|
|
goto error;
|
|
|
|
Py_DECREF(pickler);
|
|
Py_RETURN_NONE;
|
|
|
|
error:
|
|
Py_XDECREF(pickler);
|
|
return NULL;
|
|
}
|
|
|
|
/*[clinic input]
|
|
|
|
_pickle.dumps
|
|
|
|
obj: object
|
|
protocol: object = None
|
|
*
|
|
fix_imports: bool = True
|
|
buffer_callback: object = None
|
|
|
|
Return the pickled representation of the object as a bytes object.
|
|
|
|
The optional *protocol* argument tells the pickler to use the given
|
|
protocol; supported protocols are 0, 1, 2, 3, 4 and 5. The default
|
|
protocol is 4. It was introduced in Python 3.4, and is incompatible
|
|
with previous versions.
|
|
|
|
Specifying a negative protocol version selects the highest protocol
|
|
version supported. The higher the protocol used, the more recent the
|
|
version of Python needed to read the pickle produced.
|
|
|
|
If *fix_imports* is True and *protocol* is less than 3, pickle will
|
|
try to map the new Python 3 names to the old module names used in
|
|
Python 2, so that the pickle data stream is readable with Python 2.
|
|
|
|
If *buffer_callback* is None (the default), buffer views are serialized
|
|
into *file* as part of the pickle stream. It is an error if
|
|
*buffer_callback* is not None and *protocol* is None or smaller than 5.
|
|
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_dumps_impl(PyObject *module, PyObject *obj, PyObject *protocol,
|
|
int fix_imports, PyObject *buffer_callback)
|
|
/*[clinic end generated code: output=fbab0093a5580fdf input=e543272436c6f987]*/
|
|
{
|
|
PyObject *result;
|
|
PicklerObject *pickler = _Pickler_New();
|
|
|
|
if (pickler == NULL)
|
|
return NULL;
|
|
|
|
if (_Pickler_SetProtocol(pickler, protocol, fix_imports) < 0)
|
|
goto error;
|
|
|
|
if (_Pickler_SetBufferCallback(pickler, buffer_callback) < 0)
|
|
goto error;
|
|
|
|
if (dump(pickler, obj) < 0)
|
|
goto error;
|
|
|
|
result = _Pickler_GetString(pickler);
|
|
Py_DECREF(pickler);
|
|
return result;
|
|
|
|
error:
|
|
Py_XDECREF(pickler);
|
|
return NULL;
|
|
}
|
|
|
|
/*[clinic input]
|
|
|
|
_pickle.load
|
|
|
|
file: object
|
|
*
|
|
fix_imports: bool = True
|
|
encoding: str = 'ASCII'
|
|
errors: str = 'strict'
|
|
buffers: object(c_default="NULL") = ()
|
|
|
|
Read and return an object from the pickle data stored in a file.
|
|
|
|
This is equivalent to ``Unpickler(file).load()``, but may be more
|
|
efficient.
|
|
|
|
The protocol version of the pickle is detected automatically, so no
|
|
protocol argument is needed. Bytes past the pickled object's
|
|
representation are ignored.
|
|
|
|
The argument *file* must have two methods, a read() method that takes
|
|
an integer argument, and a readline() method that requires no
|
|
arguments. Both methods should return bytes. Thus *file* can be a
|
|
binary file object opened for reading, an io.BytesIO object, or any
|
|
other custom object that meets this interface.
|
|
|
|
Optional keyword arguments are *fix_imports*, *encoding* and *errors*,
|
|
which are used to control compatibility support for pickle stream
|
|
generated by Python 2. If *fix_imports* is True, pickle will try to
|
|
map the old Python 2 names to the new names used in Python 3. The
|
|
*encoding* and *errors* tell pickle how to decode 8-bit string
|
|
instances pickled by Python 2; these default to 'ASCII' and 'strict',
|
|
respectively. The *encoding* can be 'bytes' to read these 8-bit
|
|
string instances as bytes objects.
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_load_impl(PyObject *module, PyObject *file, int fix_imports,
|
|
const char *encoding, const char *errors,
|
|
PyObject *buffers)
|
|
/*[clinic end generated code: output=250452d141c23e76 input=46c7c31c92f4f371]*/
|
|
{
|
|
PyObject *result;
|
|
UnpicklerObject *unpickler = _Unpickler_New();
|
|
|
|
if (unpickler == NULL)
|
|
return NULL;
|
|
|
|
if (_Unpickler_SetInputStream(unpickler, file) < 0)
|
|
goto error;
|
|
|
|
if (_Unpickler_SetInputEncoding(unpickler, encoding, errors) < 0)
|
|
goto error;
|
|
|
|
if (_Unpickler_SetBuffers(unpickler, buffers) < 0)
|
|
goto error;
|
|
|
|
unpickler->fix_imports = fix_imports;
|
|
|
|
result = load(unpickler);
|
|
Py_DECREF(unpickler);
|
|
return result;
|
|
|
|
error:
|
|
Py_XDECREF(unpickler);
|
|
return NULL;
|
|
}
|
|
|
|
/*[clinic input]
|
|
|
|
_pickle.loads
|
|
|
|
data: object
|
|
/
|
|
*
|
|
fix_imports: bool = True
|
|
encoding: str = 'ASCII'
|
|
errors: str = 'strict'
|
|
buffers: object(c_default="NULL") = ()
|
|
|
|
Read and return an object from the given pickle data.
|
|
|
|
The protocol version of the pickle is detected automatically, so no
|
|
protocol argument is needed. Bytes past the pickled object's
|
|
representation are ignored.
|
|
|
|
Optional keyword arguments are *fix_imports*, *encoding* and *errors*,
|
|
which are used to control compatibility support for pickle stream
|
|
generated by Python 2. If *fix_imports* is True, pickle will try to
|
|
map the old Python 2 names to the new names used in Python 3. The
|
|
*encoding* and *errors* tell pickle how to decode 8-bit string
|
|
instances pickled by Python 2; these default to 'ASCII' and 'strict',
|
|
respectively. The *encoding* can be 'bytes' to read these 8-bit
|
|
string instances as bytes objects.
|
|
[clinic start generated code]*/
|
|
|
|
static PyObject *
|
|
_pickle_loads_impl(PyObject *module, PyObject *data, int fix_imports,
|
|
const char *encoding, const char *errors,
|
|
PyObject *buffers)
|
|
/*[clinic end generated code: output=82ac1e6b588e6d02 input=b3615540d0535087]*/
|
|
{
|
|
PyObject *result;
|
|
UnpicklerObject *unpickler = _Unpickler_New();
|
|
|
|
if (unpickler == NULL)
|
|
return NULL;
|
|
|
|
if (_Unpickler_SetStringInput(unpickler, data) < 0)
|
|
goto error;
|
|
|
|
if (_Unpickler_SetInputEncoding(unpickler, encoding, errors) < 0)
|
|
goto error;
|
|
|
|
if (_Unpickler_SetBuffers(unpickler, buffers) < 0)
|
|
goto error;
|
|
|
|
unpickler->fix_imports = fix_imports;
|
|
|
|
result = load(unpickler);
|
|
Py_DECREF(unpickler);
|
|
return result;
|
|
|
|
error:
|
|
Py_XDECREF(unpickler);
|
|
return NULL;
|
|
}
|
|
|
|
static struct PyMethodDef pickle_methods[] = {
|
|
_PICKLE_DUMP_METHODDEF
|
|
_PICKLE_DUMPS_METHODDEF
|
|
_PICKLE_LOAD_METHODDEF
|
|
_PICKLE_LOADS_METHODDEF
|
|
{NULL, NULL} /* sentinel */
|
|
};
|
|
|
|
static int
|
|
pickle_clear(PyObject *m)
|
|
{
|
|
_Pickle_ClearState(_Pickle_GetState(m));
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
pickle_free(PyObject *m)
|
|
{
|
|
_Pickle_ClearState(_Pickle_GetState(m));
|
|
}
|
|
|
|
static int
|
|
pickle_traverse(PyObject *m, visitproc visit, void *arg)
|
|
{
|
|
PickleState *st = _Pickle_GetState(m);
|
|
Py_VISIT(st->PickleError);
|
|
Py_VISIT(st->PicklingError);
|
|
Py_VISIT(st->UnpicklingError);
|
|
Py_VISIT(st->dispatch_table);
|
|
Py_VISIT(st->extension_registry);
|
|
Py_VISIT(st->extension_cache);
|
|
Py_VISIT(st->inverted_registry);
|
|
Py_VISIT(st->name_mapping_2to3);
|
|
Py_VISIT(st->import_mapping_2to3);
|
|
Py_VISIT(st->name_mapping_3to2);
|
|
Py_VISIT(st->import_mapping_3to2);
|
|
Py_VISIT(st->codecs_encode);
|
|
Py_VISIT(st->getattr);
|
|
Py_VISIT(st->partial);
|
|
return 0;
|
|
}
|
|
|
|
static struct PyModuleDef _picklemodule = {
|
|
PyModuleDef_HEAD_INIT,
|
|
"_pickle", /* m_name */
|
|
pickle_module_doc, /* m_doc */
|
|
sizeof(PickleState), /* m_size */
|
|
pickle_methods, /* m_methods */
|
|
NULL, /* m_reload */
|
|
pickle_traverse, /* m_traverse */
|
|
pickle_clear, /* m_clear */
|
|
(freefunc)pickle_free /* m_free */
|
|
};
|
|
|
|
PyMODINIT_FUNC
|
|
PyInit__pickle(void)
|
|
{
|
|
PyObject *m;
|
|
PickleState *st;
|
|
|
|
m = PyState_FindModule(&_picklemodule);
|
|
if (m) {
|
|
Py_INCREF(m);
|
|
return m;
|
|
}
|
|
|
|
if (PyType_Ready(&Pdata_Type) < 0)
|
|
return NULL;
|
|
if (PyType_Ready(&PicklerMemoProxyType) < 0)
|
|
return NULL;
|
|
if (PyType_Ready(&UnpicklerMemoProxyType) < 0)
|
|
return NULL;
|
|
|
|
/* Create the module and add the functions. */
|
|
m = PyModule_Create(&_picklemodule);
|
|
if (m == NULL)
|
|
return NULL;
|
|
|
|
/* Add types */
|
|
if (PyModule_AddType(m, &Pickler_Type) < 0) {
|
|
return NULL;
|
|
}
|
|
if (PyModule_AddType(m, &Unpickler_Type) < 0) {
|
|
return NULL;
|
|
}
|
|
if (PyModule_AddType(m, &PyPickleBuffer_Type) < 0) {
|
|
return NULL;
|
|
}
|
|
|
|
st = _Pickle_GetState(m);
|
|
|
|
/* Initialize the exceptions. */
|
|
st->PickleError = PyErr_NewException("_pickle.PickleError", NULL, NULL);
|
|
if (st->PickleError == NULL)
|
|
return NULL;
|
|
st->PicklingError = \
|
|
PyErr_NewException("_pickle.PicklingError", st->PickleError, NULL);
|
|
if (st->PicklingError == NULL)
|
|
return NULL;
|
|
st->UnpicklingError = \
|
|
PyErr_NewException("_pickle.UnpicklingError", st->PickleError, NULL);
|
|
if (st->UnpicklingError == NULL)
|
|
return NULL;
|
|
|
|
Py_INCREF(st->PickleError);
|
|
if (PyModule_AddObject(m, "PickleError", st->PickleError) < 0)
|
|
return NULL;
|
|
Py_INCREF(st->PicklingError);
|
|
if (PyModule_AddObject(m, "PicklingError", st->PicklingError) < 0)
|
|
return NULL;
|
|
Py_INCREF(st->UnpicklingError);
|
|
if (PyModule_AddObject(m, "UnpicklingError", st->UnpicklingError) < 0)
|
|
return NULL;
|
|
|
|
if (_Pickle_InitState(st) < 0)
|
|
return NULL;
|
|
|
|
return m;
|
|
}
|