2019-06-03 09:30:58 -03:00
|
|
|
#ifndef Py_INTERNAL_CODE_H
|
|
|
|
#define Py_INTERNAL_CODE_H
|
|
|
|
#ifdef __cplusplus
|
|
|
|
extern "C" {
|
|
|
|
#endif
|
2021-05-27 12:54:34 -03:00
|
|
|
|
2023-08-21 14:15:52 -03:00
|
|
|
#ifndef Py_BUILD_CORE
|
|
|
|
# error "this header requires Py_BUILD_CORE define"
|
|
|
|
#endif
|
|
|
|
|
2024-06-26 16:10:43 -03:00
|
|
|
#include "pycore_stackref.h" // _PyStackRef
|
2024-05-06 21:12:39 -03:00
|
|
|
#include "pycore_lock.h" // PyMutex
|
2024-06-26 08:54:03 -03:00
|
|
|
#include "pycore_backoff.h" // _Py_BackoffCounter
|
2024-11-04 15:13:32 -04:00
|
|
|
#include "pycore_tstate.h" // _PyThreadStateImpl
|
2024-06-26 08:54:03 -03:00
|
|
|
|
|
|
|
|
|
|
|
/* Each instruction in a code object is a fixed-width value,
|
|
|
|
* currently 2 bytes: 1-byte opcode + 1-byte oparg. The EXTENDED_ARG
|
|
|
|
* opcode allows for larger values but the current limit is 3 uses
|
|
|
|
* of EXTENDED_ARG (see Python/compile.c), for a maximum
|
|
|
|
* 32-bit value. This aligns with the note in Python/compile.c
|
|
|
|
* (compiler_addop_i_line) indicating that the max oparg value is
|
|
|
|
* 2**32 - 1, rather than INT_MAX.
|
|
|
|
*/
|
|
|
|
|
|
|
|
typedef union {
|
|
|
|
uint16_t cache;
|
|
|
|
struct {
|
|
|
|
uint8_t code;
|
|
|
|
uint8_t arg;
|
|
|
|
} op;
|
|
|
|
_Py_BackoffCounter counter; // First cache entry of specializable op
|
|
|
|
} _Py_CODEUNIT;
|
|
|
|
|
2024-07-13 18:07:49 -03:00
|
|
|
#define _PyCode_CODE(CO) _Py_RVALUE((_Py_CODEUNIT *)(CO)->co_code_adaptive)
|
|
|
|
#define _PyCode_NBYTES(CO) (Py_SIZE(CO) * (Py_ssize_t)sizeof(_Py_CODEUNIT))
|
|
|
|
|
2024-06-26 08:54:03 -03:00
|
|
|
|
|
|
|
/* These macros only remain defined for compatibility. */
|
|
|
|
#define _Py_OPCODE(word) ((word).op.code)
|
|
|
|
#define _Py_OPARG(word) ((word).op.arg)
|
|
|
|
|
|
|
|
static inline _Py_CODEUNIT
|
|
|
|
_py_make_codeunit(uint8_t opcode, uint8_t oparg)
|
|
|
|
{
|
|
|
|
// No designated initialisers because of C++ compat
|
|
|
|
_Py_CODEUNIT word;
|
|
|
|
word.op.code = opcode;
|
|
|
|
word.op.arg = oparg;
|
|
|
|
return word;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
_py_set_opcode(_Py_CODEUNIT *word, uint8_t opcode)
|
|
|
|
{
|
|
|
|
word->op.code = opcode;
|
|
|
|
}
|
|
|
|
|
|
|
|
#define _Py_MAKE_CODEUNIT(opcode, oparg) _py_make_codeunit((opcode), (oparg))
|
|
|
|
#define _Py_SET_OPCODE(word, opcode) _py_set_opcode(&(word), (opcode))
|
2024-05-06 21:12:39 -03:00
|
|
|
|
2024-02-13 17:56:49 -04:00
|
|
|
|
|
|
|
// We hide some of the newer PyCodeObject fields behind macros.
|
|
|
|
// This helps with backporting certain changes to 3.12.
|
|
|
|
#define _PyCode_HAS_EXECUTORS(CODE) \
|
|
|
|
(CODE->co_executors != NULL)
|
|
|
|
#define _PyCode_HAS_INSTRUMENTATION(CODE) \
|
|
|
|
(CODE->_co_instrumentation_version > 0)
|
|
|
|
|
2024-05-06 21:12:39 -03:00
|
|
|
struct _py_code_state {
|
|
|
|
PyMutex mutex;
|
|
|
|
// Interned constants from code objects. Used by the free-threaded build.
|
|
|
|
struct _Py_hashtable_t *constants;
|
|
|
|
};
|
|
|
|
|
|
|
|
extern PyStatus _PyCode_Init(PyInterpreterState *interp);
|
|
|
|
extern void _PyCode_Fini(PyInterpreterState *interp);
|
2024-02-13 17:56:49 -04:00
|
|
|
|
2022-12-02 13:28:27 -04:00
|
|
|
#define CODE_MAX_WATCHERS 8
|
|
|
|
|
2021-06-07 14:38:06 -03:00
|
|
|
/* PEP 659
|
|
|
|
* Specialization and quickening structs and helper functions
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
2022-03-07 15:45:00 -04:00
|
|
|
// Inline caches. If you change the number of cache entries for an instruction,
|
|
|
|
// you must *also* update the number of cache entries in Lib/opcode.py and bump
|
|
|
|
// the magic number in Lib/importlib/_bootstrap_external.py!
|
2022-02-28 08:56:29 -04:00
|
|
|
|
|
|
|
#define CACHE_ENTRIES(cache) (sizeof(cache)/sizeof(_Py_CODEUNIT))
|
|
|
|
|
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2023-03-10 21:01:16 -04:00
|
|
|
uint16_t module_keys_version;
|
2022-12-14 07:12:53 -04:00
|
|
|
uint16_t builtin_keys_version;
|
2023-07-12 10:34:14 -03:00
|
|
|
uint16_t index;
|
2022-02-28 08:56:29 -04:00
|
|
|
} _PyLoadGlobalCache;
|
|
|
|
|
|
|
|
#define INLINE_CACHE_ENTRIES_LOAD_GLOBAL CACHE_ENTRIES(_PyLoadGlobalCache)
|
|
|
|
|
2022-02-25 08:11:34 -04:00
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2022-02-25 08:11:34 -04:00
|
|
|
} _PyBinaryOpCache;
|
|
|
|
|
2022-02-28 08:56:29 -04:00
|
|
|
#define INLINE_CACHE_ENTRIES_BINARY_OP CACHE_ENTRIES(_PyBinaryOpCache)
|
2022-03-01 09:53:13 -04:00
|
|
|
|
2022-02-28 07:54:14 -04:00
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2022-02-28 07:54:14 -04:00
|
|
|
} _PyUnpackSequenceCache;
|
|
|
|
|
|
|
|
#define INLINE_CACHE_ENTRIES_UNPACK_SEQUENCE \
|
2022-03-01 09:53:13 -04:00
|
|
|
CACHE_ENTRIES(_PyUnpackSequenceCache)
|
|
|
|
|
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2022-03-01 09:53:13 -04:00
|
|
|
} _PyCompareOpCache;
|
|
|
|
|
|
|
|
#define INLINE_CACHE_ENTRIES_COMPARE_OP CACHE_ENTRIES(_PyCompareOpCache)
|
2022-02-28 07:54:14 -04:00
|
|
|
|
2022-03-01 12:00:34 -04:00
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2022-03-01 12:00:34 -04:00
|
|
|
} _PyBinarySubscrCache;
|
|
|
|
|
|
|
|
#define INLINE_CACHE_ENTRIES_BINARY_SUBSCR CACHE_ENTRIES(_PyBinarySubscrCache)
|
|
|
|
|
2023-04-25 14:45:51 -03:00
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2023-04-25 14:45:51 -03:00
|
|
|
} _PySuperAttrCache;
|
|
|
|
|
|
|
|
#define INLINE_CACHE_ENTRIES_LOAD_SUPER_ATTR CACHE_ENTRIES(_PySuperAttrCache)
|
|
|
|
|
2022-03-03 19:31:00 -04:00
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2022-12-14 07:12:53 -04:00
|
|
|
uint16_t version[2];
|
|
|
|
uint16_t index;
|
2022-03-03 19:31:00 -04:00
|
|
|
} _PyAttrCache;
|
|
|
|
|
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2022-12-14 07:12:53 -04:00
|
|
|
uint16_t type_version[2];
|
2024-04-02 07:59:21 -03:00
|
|
|
union {
|
|
|
|
uint16_t keys_version[2];
|
|
|
|
uint16_t dict_offset;
|
|
|
|
};
|
2022-12-14 07:12:53 -04:00
|
|
|
uint16_t descr[4];
|
2022-03-03 19:31:00 -04:00
|
|
|
} _PyLoadMethodCache;
|
|
|
|
|
2022-06-14 07:36:22 -03:00
|
|
|
|
|
|
|
// MUST be the max(_PyAttrCache, _PyLoadMethodCache)
|
|
|
|
#define INLINE_CACHE_ENTRIES_LOAD_ATTR CACHE_ENTRIES(_PyLoadMethodCache)
|
|
|
|
|
|
|
|
#define INLINE_CACHE_ENTRIES_STORE_ATTR CACHE_ENTRIES(_PyAttrCache)
|
2022-03-03 19:31:00 -04:00
|
|
|
|
2022-03-07 15:45:00 -04:00
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2022-12-14 07:12:53 -04:00
|
|
|
uint16_t func_version[2];
|
2022-03-07 15:45:00 -04:00
|
|
|
} _PyCallCache;
|
2021-06-07 14:38:06 -03:00
|
|
|
|
2022-03-07 15:45:00 -04:00
|
|
|
#define INLINE_CACHE_ENTRIES_CALL CACHE_ENTRIES(_PyCallCache)
|
2024-08-16 13:11:24 -03:00
|
|
|
#define INLINE_CACHE_ENTRIES_CALL_KW CACHE_ENTRIES(_PyCallCache)
|
2021-06-07 14:38:06 -03:00
|
|
|
|
2022-03-08 11:53:22 -04:00
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2022-03-08 11:53:22 -04:00
|
|
|
} _PyStoreSubscrCache;
|
|
|
|
|
|
|
|
#define INLINE_CACHE_ENTRIES_STORE_SUBSCR CACHE_ENTRIES(_PyStoreSubscrCache)
|
|
|
|
|
2022-06-21 07:19:26 -03:00
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2022-06-21 07:19:26 -03:00
|
|
|
} _PyForIterCache;
|
|
|
|
|
|
|
|
#define INLINE_CACHE_ENTRIES_FOR_ITER CACHE_ENTRIES(_PyForIterCache)
|
|
|
|
|
2023-02-13 07:24:55 -04:00
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2023-02-13 07:24:55 -04:00
|
|
|
} _PySendCache;
|
|
|
|
|
|
|
|
#define INLINE_CACHE_ENTRIES_SEND CACHE_ENTRIES(_PySendCache)
|
|
|
|
|
2023-06-29 17:49:54 -03:00
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2023-06-29 17:49:54 -03:00
|
|
|
uint16_t version[2];
|
|
|
|
} _PyToBoolCache;
|
|
|
|
|
|
|
|
#define INLINE_CACHE_ENTRIES_TO_BOOL CACHE_ENTRIES(_PyToBoolCache)
|
|
|
|
|
2024-03-06 15:30:11 -04:00
|
|
|
typedef struct {
|
2024-04-04 12:03:27 -03:00
|
|
|
_Py_BackoffCounter counter;
|
2024-03-06 15:30:11 -04:00
|
|
|
} _PyContainsOpCache;
|
|
|
|
|
|
|
|
#define INLINE_CACHE_ENTRIES_CONTAINS_OP CACHE_ENTRIES(_PyContainsOpCache)
|
|
|
|
|
2022-03-07 15:45:00 -04:00
|
|
|
// Borrowed references to common callables:
|
|
|
|
struct callable_cache {
|
|
|
|
PyObject *isinstance;
|
|
|
|
PyObject *len;
|
|
|
|
PyObject *list_append;
|
2022-08-17 08:37:07 -03:00
|
|
|
PyObject *object__getattribute__;
|
2022-03-07 15:45:00 -04:00
|
|
|
};
|
|
|
|
|
2021-06-07 15:22:26 -03:00
|
|
|
/* "Locals plus" for a code object is the set of locals + cell vars +
|
|
|
|
* free vars. This relates to variable names as well as offsets into
|
|
|
|
* the "fast locals" storage array of execution frames. The compiler
|
|
|
|
* builds the list of names, their offsets, and the corresponding
|
|
|
|
* kind of local.
|
|
|
|
*
|
|
|
|
* Those kinds represent the source of the initial value and the
|
|
|
|
* variable's scope (as related to closures). A "local" is an
|
|
|
|
* argument or other variable defined in the current scope. A "free"
|
|
|
|
* variable is one that is defined in an outer scope and comes from
|
|
|
|
* the function's closure. A "cell" variable is a local that escapes
|
|
|
|
* into an inner function as part of a closure, and thus must be
|
|
|
|
* wrapped in a cell. Any "local" can also be a "cell", but the
|
|
|
|
* "free" kind is mutually exclusive with both.
|
|
|
|
*/
|
|
|
|
|
2021-06-21 17:53:04 -03:00
|
|
|
// Note that these all fit within a byte, as do combinations.
|
2021-06-07 15:22:26 -03:00
|
|
|
// Later, we will use the smaller numbers to differentiate the different
|
|
|
|
// kinds of locals (e.g. pos-only arg, varkwargs, local-only).
|
2023-05-09 14:02:14 -03:00
|
|
|
#define CO_FAST_HIDDEN 0x10
|
2021-06-07 15:22:26 -03:00
|
|
|
#define CO_FAST_LOCAL 0x20
|
|
|
|
#define CO_FAST_CELL 0x40
|
|
|
|
#define CO_FAST_FREE 0x80
|
|
|
|
|
2021-06-21 17:53:04 -03:00
|
|
|
typedef unsigned char _PyLocals_Kind;
|
|
|
|
|
|
|
|
static inline _PyLocals_Kind
|
|
|
|
_PyLocals_GetKind(PyObject *kinds, int i)
|
2021-06-07 15:22:26 -03:00
|
|
|
{
|
2021-06-21 17:53:04 -03:00
|
|
|
assert(PyBytes_Check(kinds));
|
|
|
|
assert(0 <= i && i < PyBytes_GET_SIZE(kinds));
|
|
|
|
char *ptr = PyBytes_AS_STRING(kinds);
|
|
|
|
return (_PyLocals_Kind)(ptr[i]);
|
2021-06-07 15:22:26 -03:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
2021-06-21 17:53:04 -03:00
|
|
|
_PyLocals_SetKind(PyObject *kinds, int i, _PyLocals_Kind kind)
|
2021-06-07 15:22:26 -03:00
|
|
|
{
|
2021-06-21 17:53:04 -03:00
|
|
|
assert(PyBytes_Check(kinds));
|
|
|
|
assert(0 <= i && i < PyBytes_GET_SIZE(kinds));
|
|
|
|
char *ptr = PyBytes_AS_STRING(kinds);
|
|
|
|
ptr[i] = (char) kind;
|
2021-06-07 15:22:26 -03:00
|
|
|
}
|
|
|
|
|
2021-06-21 17:53:04 -03:00
|
|
|
|
2021-05-27 12:54:34 -03:00
|
|
|
struct _PyCodeConstructor {
|
|
|
|
/* metadata */
|
|
|
|
PyObject *filename;
|
|
|
|
PyObject *name;
|
2021-07-07 08:21:51 -03:00
|
|
|
PyObject *qualname;
|
2021-05-27 12:54:34 -03:00
|
|
|
int flags;
|
|
|
|
|
|
|
|
/* the code */
|
|
|
|
PyObject *code;
|
|
|
|
int firstlineno;
|
|
|
|
PyObject *linetable;
|
|
|
|
|
|
|
|
/* used by the code */
|
|
|
|
PyObject *consts;
|
|
|
|
PyObject *names;
|
|
|
|
|
|
|
|
/* mapping frame offsets to information */
|
2021-06-21 17:53:04 -03:00
|
|
|
PyObject *localsplusnames; // Tuple of strings
|
|
|
|
PyObject *localspluskinds; // Bytes object, one byte per variable
|
2021-05-27 12:54:34 -03:00
|
|
|
|
|
|
|
/* args (within varnames) */
|
|
|
|
int argcount;
|
|
|
|
int posonlyargcount;
|
2021-06-07 15:22:26 -03:00
|
|
|
// XXX Replace argcount with posorkwargcount (argcount - posonlyargcount).
|
2021-05-27 12:54:34 -03:00
|
|
|
int kwonlyargcount;
|
|
|
|
|
|
|
|
/* needed to create the frame */
|
|
|
|
int stacksize;
|
|
|
|
|
|
|
|
/* used by the eval loop */
|
|
|
|
PyObject *exceptiontable;
|
|
|
|
};
|
|
|
|
|
|
|
|
// Using an "arguments struct" like this is helpful for maintainability
|
|
|
|
// in a case such as this with many parameters. It does bear a risk:
|
|
|
|
// if the struct changes and callers are not updated properly then the
|
|
|
|
// compiler will not catch problems (like a missing argument). This can
|
|
|
|
// cause hard-to-debug problems. The risk is mitigated by the use of
|
|
|
|
// check_code() in codeobject.c. However, we may decide to switch
|
|
|
|
// back to a regular function signature. Regardless, this approach
|
|
|
|
// wouldn't be appropriate if this weren't a strictly internal API.
|
|
|
|
// (See the comments in https://github.com/python/cpython/pull/26258.)
|
2023-07-24 22:44:11 -03:00
|
|
|
extern int _PyCode_Validate(struct _PyCodeConstructor *);
|
|
|
|
extern PyCodeObject* _PyCode_New(struct _PyCodeConstructor *);
|
2021-05-27 12:54:34 -03:00
|
|
|
|
|
|
|
|
2019-06-03 09:30:58 -03:00
|
|
|
/* Private API */
|
2021-05-27 12:54:34 -03:00
|
|
|
|
2021-06-07 15:22:26 -03:00
|
|
|
/* Getters for internal PyCodeObject data. */
|
2022-02-25 10:41:32 -04:00
|
|
|
extern PyObject* _PyCode_GetVarnames(PyCodeObject *);
|
|
|
|
extern PyObject* _PyCode_GetCellvars(PyCodeObject *);
|
|
|
|
extern PyObject* _PyCode_GetFreevars(PyCodeObject *);
|
2022-03-21 08:11:17 -03:00
|
|
|
extern PyObject* _PyCode_GetCode(PyCodeObject *);
|
2022-02-25 10:41:32 -04:00
|
|
|
|
|
|
|
/** API for initializing the line number tables. */
|
|
|
|
extern int _PyCode_InitAddressRange(PyCodeObject* co, PyCodeAddressRange *bounds);
|
|
|
|
|
2022-04-21 12:10:37 -03:00
|
|
|
/** Out of process API for initializing the location table. */
|
2022-02-25 10:41:32 -04:00
|
|
|
extern void _PyLineTable_InitAddressRange(
|
|
|
|
const char *linetable,
|
|
|
|
Py_ssize_t length,
|
|
|
|
int firstlineno,
|
|
|
|
PyCodeAddressRange *range);
|
|
|
|
|
|
|
|
/** API for traversing the line number table. */
|
|
|
|
extern int _PyLineTable_NextAddressRange(PyCodeAddressRange *range);
|
|
|
|
extern int _PyLineTable_PreviousAddressRange(PyCodeAddressRange *range);
|
|
|
|
|
2023-09-06 21:53:54 -03:00
|
|
|
/** API for executors */
|
|
|
|
extern void _PyCode_Clear_Executors(PyCodeObject *code);
|
|
|
|
|
2024-11-04 15:13:32 -04:00
|
|
|
|
2024-02-29 22:53:32 -04:00
|
|
|
#ifdef Py_GIL_DISABLED
|
|
|
|
// gh-115999 tracks progress on addressing this.
|
|
|
|
#define ENABLE_SPECIALIZATION 0
|
2024-11-04 15:13:32 -04:00
|
|
|
// Use this to enable specialization families once they are thread-safe. All
|
|
|
|
// uses will be replaced with ENABLE_SPECIALIZATION once all families are
|
|
|
|
// thread-safe.
|
|
|
|
#define ENABLE_SPECIALIZATION_FT 1
|
2024-02-29 22:53:32 -04:00
|
|
|
#else
|
2023-08-01 14:05:00 -03:00
|
|
|
#define ENABLE_SPECIALIZATION 1
|
2024-11-04 15:13:32 -04:00
|
|
|
#define ENABLE_SPECIALIZATION_FT ENABLE_SPECIALIZATION
|
2024-02-29 22:53:32 -04:00
|
|
|
#endif
|
2023-08-01 14:05:00 -03:00
|
|
|
|
2021-06-10 04:46:01 -03:00
|
|
|
/* Specialization functions */
|
|
|
|
|
2024-06-26 16:10:43 -03:00
|
|
|
extern void _Py_Specialize_LoadSuperAttr(_PyStackRef global_super, _PyStackRef cls,
|
2023-05-11 11:08:13 -03:00
|
|
|
_Py_CODEUNIT *instr, int load_method);
|
2024-06-26 16:10:43 -03:00
|
|
|
extern void _Py_Specialize_LoadAttr(_PyStackRef owner, _Py_CODEUNIT *instr,
|
2022-03-03 19:31:00 -04:00
|
|
|
PyObject *name);
|
2024-06-26 16:10:43 -03:00
|
|
|
extern void _Py_Specialize_StoreAttr(_PyStackRef owner, _Py_CODEUNIT *instr,
|
2022-11-17 19:09:18 -04:00
|
|
|
PyObject *name);
|
2022-11-09 14:50:09 -04:00
|
|
|
extern void _Py_Specialize_LoadGlobal(PyObject *globals, PyObject *builtins,
|
|
|
|
_Py_CODEUNIT *instr, PyObject *name);
|
2024-06-26 16:10:43 -03:00
|
|
|
extern void _Py_Specialize_BinarySubscr(_PyStackRef sub, _PyStackRef container,
|
2022-11-09 14:50:09 -04:00
|
|
|
_Py_CODEUNIT *instr);
|
2024-06-26 16:10:43 -03:00
|
|
|
extern void _Py_Specialize_StoreSubscr(_PyStackRef container, _PyStackRef sub,
|
2022-11-09 14:50:09 -04:00
|
|
|
_Py_CODEUNIT *instr);
|
2024-06-26 16:10:43 -03:00
|
|
|
extern void _Py_Specialize_Call(_PyStackRef callable, _Py_CODEUNIT *instr,
|
2023-09-13 14:25:45 -03:00
|
|
|
int nargs);
|
2024-08-16 13:11:24 -03:00
|
|
|
extern void _Py_Specialize_CallKw(_PyStackRef callable, _Py_CODEUNIT *instr,
|
|
|
|
int nargs);
|
2024-06-26 16:10:43 -03:00
|
|
|
extern void _Py_Specialize_BinaryOp(_PyStackRef lhs, _PyStackRef rhs, _Py_CODEUNIT *instr,
|
|
|
|
int oparg, _PyStackRef *locals);
|
|
|
|
extern void _Py_Specialize_CompareOp(_PyStackRef lhs, _PyStackRef rhs,
|
2022-03-01 09:53:13 -04:00
|
|
|
_Py_CODEUNIT *instr, int oparg);
|
2024-06-26 16:10:43 -03:00
|
|
|
extern void _Py_Specialize_UnpackSequence(_PyStackRef seq, _Py_CODEUNIT *instr,
|
2022-02-28 07:54:14 -04:00
|
|
|
int oparg);
|
2024-06-26 16:10:43 -03:00
|
|
|
extern void _Py_Specialize_ForIter(_PyStackRef iter, _Py_CODEUNIT *instr, int oparg);
|
|
|
|
extern void _Py_Specialize_Send(_PyStackRef receiver, _Py_CODEUNIT *instr);
|
|
|
|
extern void _Py_Specialize_ToBool(_PyStackRef value, _Py_CODEUNIT *instr);
|
|
|
|
extern void _Py_Specialize_ContainsOp(_PyStackRef value, _Py_CODEUNIT *instr);
|
2021-06-10 04:46:01 -03:00
|
|
|
|
2021-12-15 11:32:32 -04:00
|
|
|
#ifdef Py_STATS
|
2021-06-15 09:01:42 -03:00
|
|
|
|
2023-11-03 12:21:16 -03:00
|
|
|
#include "pycore_bitutils.h" // _Py_bit_length
|
|
|
|
|
2023-09-06 12:54:59 -03:00
|
|
|
#define STAT_INC(opname, name) do { if (_Py_stats) _Py_stats->opcode_stats[opname].specialization.name++; } while (0)
|
|
|
|
#define STAT_DEC(opname, name) do { if (_Py_stats) _Py_stats->opcode_stats[opname].specialization.name--; } while (0)
|
|
|
|
#define OPCODE_EXE_INC(opname) do { if (_Py_stats) _Py_stats->opcode_stats[opname].execution_count++; } while (0)
|
|
|
|
#define CALL_STAT_INC(name) do { if (_Py_stats) _Py_stats->call_stats.name++; } while (0)
|
|
|
|
#define OBJECT_STAT_INC(name) do { if (_Py_stats) _Py_stats->object_stats.name++; } while (0)
|
2022-05-03 19:40:24 -03:00
|
|
|
#define OBJECT_STAT_INC_COND(name, cond) \
|
2023-09-06 12:54:59 -03:00
|
|
|
do { if (_Py_stats && cond) _Py_stats->object_stats.name++; } while (0)
|
|
|
|
#define EVAL_CALL_STAT_INC(name) do { if (_Py_stats) _Py_stats->call_stats.eval_calls[name]++; } while (0)
|
2022-05-27 12:31:41 -03:00
|
|
|
#define EVAL_CALL_STAT_INC_IF_FUNCTION(name, callable) \
|
2023-09-06 12:54:59 -03:00
|
|
|
do { if (_Py_stats && PyFunction_Check(callable)) _Py_stats->call_stats.eval_calls[name]++; } while (0)
|
|
|
|
#define GC_STAT_ADD(gen, name, n) do { if (_Py_stats) _Py_stats->gc_stats[(gen)].name += (n); } while (0)
|
2023-10-04 18:52:28 -03:00
|
|
|
#define OPT_STAT_INC(name) do { if (_Py_stats) _Py_stats->optimization_stats.name++; } while (0)
|
2023-10-30 21:02:45 -03:00
|
|
|
#define UOP_STAT_INC(opname, name) do { if (_Py_stats) { assert(opname < 512); _Py_stats->optimization_stats.opcode[opname].name++; } } while (0)
|
2024-04-16 10:27:18 -03:00
|
|
|
#define UOP_PAIR_INC(uopcode, lastuop) \
|
|
|
|
do { \
|
|
|
|
if (lastuop && _Py_stats) { \
|
|
|
|
_Py_stats->optimization_stats.opcode[lastuop].pair_count[uopcode]++; \
|
|
|
|
} \
|
|
|
|
lastuop = uopcode; \
|
|
|
|
} while (0)
|
2023-10-04 18:52:28 -03:00
|
|
|
#define OPT_UNSUPPORTED_OPCODE(opname) do { if (_Py_stats) _Py_stats->optimization_stats.unsupported_opcode[opname]++; } while (0)
|
2024-03-21 14:27:46 -03:00
|
|
|
#define OPT_ERROR_IN_OPCODE(opname) do { if (_Py_stats) _Py_stats->optimization_stats.error_in_opcode[opname]++; } while (0)
|
2023-10-04 18:52:28 -03:00
|
|
|
#define OPT_HIST(length, name) \
|
|
|
|
do { \
|
|
|
|
if (_Py_stats) { \
|
|
|
|
int bucket = _Py_bit_length(length >= 1 ? length - 1 : 0); \
|
|
|
|
bucket = (bucket >= _Py_UOP_HIST_SIZE) ? _Py_UOP_HIST_SIZE - 1 : bucket; \
|
|
|
|
_Py_stats->optimization_stats.name[bucket]++; \
|
|
|
|
} \
|
|
|
|
} while (0)
|
2024-01-25 07:10:51 -04:00
|
|
|
#define RARE_EVENT_STAT_INC(name) do { if (_Py_stats) _Py_stats->rare_event_stats.name++; } while (0)
|
2024-08-22 10:17:10 -03:00
|
|
|
#define OPCODE_DEFERRED_INC(opname) do { if (_Py_stats && opcode == opname) _Py_stats->opcode_stats[opname].specialization.deferred++; } while (0)
|
2021-12-17 10:48:01 -04:00
|
|
|
|
2023-07-25 00:16:28 -03:00
|
|
|
// Export for '_opcode' shared extension
|
2022-04-28 09:02:45 -03:00
|
|
|
PyAPI_FUNC(PyObject*) _Py_GetSpecializationStats(void);
|
2021-07-29 13:26:53 -03:00
|
|
|
|
2021-06-10 04:46:01 -03:00
|
|
|
#else
|
2021-06-14 07:04:09 -03:00
|
|
|
#define STAT_INC(opname, name) ((void)0)
|
2021-07-29 16:50:03 -03:00
|
|
|
#define STAT_DEC(opname, name) ((void)0)
|
2021-12-17 10:48:01 -04:00
|
|
|
#define OPCODE_EXE_INC(opname) ((void)0)
|
2022-01-28 11:20:33 -04:00
|
|
|
#define CALL_STAT_INC(name) ((void)0)
|
2022-02-01 11:05:18 -04:00
|
|
|
#define OBJECT_STAT_INC(name) ((void)0)
|
2022-05-03 19:40:24 -03:00
|
|
|
#define OBJECT_STAT_INC_COND(name, cond) ((void)0)
|
2022-05-27 12:31:41 -03:00
|
|
|
#define EVAL_CALL_STAT_INC(name) ((void)0)
|
|
|
|
#define EVAL_CALL_STAT_INC_IF_FUNCTION(name, callable) ((void)0)
|
2023-08-04 06:34:23 -03:00
|
|
|
#define GC_STAT_ADD(gen, name, n) ((void)0)
|
2023-10-04 18:52:28 -03:00
|
|
|
#define OPT_STAT_INC(name) ((void)0)
|
2023-10-30 21:02:45 -03:00
|
|
|
#define UOP_STAT_INC(opname, name) ((void)0)
|
2024-04-16 10:27:18 -03:00
|
|
|
#define UOP_PAIR_INC(uopcode, lastuop) ((void)0)
|
2023-10-04 18:52:28 -03:00
|
|
|
#define OPT_UNSUPPORTED_OPCODE(opname) ((void)0)
|
2024-03-21 14:27:46 -03:00
|
|
|
#define OPT_ERROR_IN_OPCODE(opname) ((void)0)
|
2023-10-04 18:52:28 -03:00
|
|
|
#define OPT_HIST(length, name) ((void)0)
|
2024-01-25 07:10:51 -04:00
|
|
|
#define RARE_EVENT_STAT_INC(name) ((void)0)
|
2024-08-22 10:17:10 -03:00
|
|
|
#define OPCODE_DEFERRED_INC(opname) ((void)0)
|
2022-04-28 09:02:45 -03:00
|
|
|
#endif // !Py_STATS
|
2021-06-10 04:46:01 -03:00
|
|
|
|
2022-09-14 21:05:04 -03:00
|
|
|
// Utility functions for reading/writing 32/64-bit values in the inline caches.
|
|
|
|
// Great care should be taken to ensure that these functions remain correct and
|
|
|
|
// performant! They should compile to just "move" instructions on all supported
|
|
|
|
// compilers and platforms.
|
|
|
|
|
|
|
|
// We use memcpy to let the C compiler handle unaligned accesses and endianness
|
|
|
|
// issues for us. It also seems to produce better code than manual copying for
|
|
|
|
// most compilers (see https://blog.regehr.org/archives/959 for more info).
|
2022-02-28 08:56:29 -04:00
|
|
|
|
|
|
|
static inline void
|
2022-03-03 19:31:00 -04:00
|
|
|
write_u32(uint16_t *p, uint32_t val)
|
2022-02-28 08:56:29 -04:00
|
|
|
{
|
2022-09-14 21:05:04 -03:00
|
|
|
memcpy(p, &val, sizeof(val));
|
2022-03-03 19:31:00 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
write_u64(uint16_t *p, uint64_t val)
|
|
|
|
{
|
2022-09-14 21:05:04 -03:00
|
|
|
memcpy(p, &val, sizeof(val));
|
2022-03-03 19:31:00 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
2022-09-14 21:05:04 -03:00
|
|
|
write_obj(uint16_t *p, PyObject *val)
|
2022-02-28 08:56:29 -04:00
|
|
|
{
|
2022-09-14 21:05:04 -03:00
|
|
|
memcpy(p, &val, sizeof(val));
|
2022-02-28 08:56:29 -04:00
|
|
|
}
|
|
|
|
|
2022-11-22 20:04:57 -04:00
|
|
|
static inline uint16_t
|
|
|
|
read_u16(uint16_t *p)
|
|
|
|
{
|
|
|
|
return *p;
|
|
|
|
}
|
|
|
|
|
2022-02-28 08:56:29 -04:00
|
|
|
static inline uint32_t
|
2022-03-03 19:31:00 -04:00
|
|
|
read_u32(uint16_t *p)
|
|
|
|
{
|
2022-09-14 21:05:04 -03:00
|
|
|
uint32_t val;
|
|
|
|
memcpy(&val, p, sizeof(val));
|
2022-03-03 19:31:00 -04:00
|
|
|
return val;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline uint64_t
|
|
|
|
read_u64(uint16_t *p)
|
|
|
|
{
|
2022-09-14 21:05:04 -03:00
|
|
|
uint64_t val;
|
|
|
|
memcpy(&val, p, sizeof(val));
|
2022-03-03 19:31:00 -04:00
|
|
|
return val;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline PyObject *
|
|
|
|
read_obj(uint16_t *p)
|
|
|
|
{
|
2022-09-14 21:05:04 -03:00
|
|
|
PyObject *val;
|
|
|
|
memcpy(&val, p, sizeof(val));
|
|
|
|
return val;
|
2022-03-03 19:31:00 -04:00
|
|
|
}
|
2021-06-10 04:46:01 -03:00
|
|
|
|
2022-07-18 12:06:42 -03:00
|
|
|
/* See Objects/exception_handling_notes.txt for details.
|
|
|
|
*/
|
|
|
|
static inline unsigned char *
|
|
|
|
parse_varint(unsigned char *p, int *result) {
|
|
|
|
int val = p[0] & 63;
|
|
|
|
while (p[0] & 64) {
|
|
|
|
p++;
|
|
|
|
val = (val << 6) | (p[0] & 63);
|
|
|
|
}
|
|
|
|
*result = val;
|
|
|
|
return p+1;
|
|
|
|
}
|
|
|
|
|
2022-04-21 12:10:37 -03:00
|
|
|
static inline int
|
|
|
|
write_varint(uint8_t *ptr, unsigned int val)
|
|
|
|
{
|
|
|
|
int written = 1;
|
|
|
|
while (val >= 64) {
|
|
|
|
*ptr++ = 64 | (val & 63);
|
|
|
|
val >>= 6;
|
|
|
|
written++;
|
|
|
|
}
|
2023-12-04 06:42:58 -04:00
|
|
|
*ptr = (uint8_t)val;
|
2022-04-21 12:10:37 -03:00
|
|
|
return written;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int
|
|
|
|
write_signed_varint(uint8_t *ptr, int val)
|
|
|
|
{
|
2023-12-04 06:42:58 -04:00
|
|
|
unsigned int uval;
|
2022-04-21 12:10:37 -03:00
|
|
|
if (val < 0) {
|
2023-12-04 06:42:58 -04:00
|
|
|
// (unsigned int)(-val) has an undefined behavior for INT_MIN
|
|
|
|
uval = ((0 - (unsigned int)val) << 1) | 1;
|
2022-04-21 12:10:37 -03:00
|
|
|
}
|
|
|
|
else {
|
2023-12-04 06:42:58 -04:00
|
|
|
uval = (unsigned int)val << 1;
|
2022-04-21 12:10:37 -03:00
|
|
|
}
|
2023-12-04 06:42:58 -04:00
|
|
|
return write_varint(ptr, uval);
|
2022-04-21 12:10:37 -03:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline int
|
|
|
|
write_location_entry_start(uint8_t *ptr, int code, int length)
|
|
|
|
{
|
|
|
|
assert((code & 15) == code);
|
2023-12-04 06:42:58 -04:00
|
|
|
*ptr = 128 | (uint8_t)(code << 3) | (uint8_t)(length - 1);
|
2022-04-21 12:10:37 -03:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2022-05-31 07:58:26 -03:00
|
|
|
/** Counters
|
|
|
|
* The first 16-bit value in each inline cache is a counter.
|
|
|
|
*
|
|
|
|
* When counting executions until the next specialization attempt,
|
|
|
|
* exponential backoff is used to reduce the number of specialization failures.
|
2024-04-04 12:03:27 -03:00
|
|
|
* See pycore_backoff.h for more details.
|
|
|
|
* On a specialization failure, the backoff counter is restarted.
|
2022-05-31 07:58:26 -03:00
|
|
|
*/
|
|
|
|
|
2024-04-04 12:03:27 -03:00
|
|
|
#include "pycore_backoff.h"
|
2022-11-09 14:50:09 -04:00
|
|
|
|
|
|
|
// A value of 1 means that we attempt to specialize the *second* time each
|
|
|
|
// instruction is executed. Executing twice is a much better indicator of
|
|
|
|
// "hotness" than executing once, but additional warmup delays only prevent
|
|
|
|
// specialization. Most types stabilize by the second execution, too:
|
|
|
|
#define ADAPTIVE_WARMUP_VALUE 1
|
|
|
|
#define ADAPTIVE_WARMUP_BACKOFF 1
|
|
|
|
|
|
|
|
// A value of 52 means that we attempt to re-specialize after 53 misses (a prime
|
|
|
|
// number, useful for avoiding artifacts if every nth value is a different type
|
|
|
|
// or something). Setting the backoff to 0 means that the counter is reset to
|
|
|
|
// the same state as a warming-up instruction (value == 1, backoff == 1) after
|
|
|
|
// deoptimization. This isn't strictly necessary, but it is bit easier to reason
|
|
|
|
// about when thinking about the opcode transitions as a state machine:
|
|
|
|
#define ADAPTIVE_COOLDOWN_VALUE 52
|
|
|
|
#define ADAPTIVE_COOLDOWN_BACKOFF 0
|
2022-05-31 07:58:26 -03:00
|
|
|
|
2024-04-04 12:03:27 -03:00
|
|
|
// Can't assert this in pycore_backoff.h because of header order dependencies
|
2024-09-06 10:52:07 -03:00
|
|
|
#if SIDE_EXIT_INITIAL_VALUE <= ADAPTIVE_COOLDOWN_VALUE
|
|
|
|
# error "Cold exit value should be larger than adaptive cooldown value"
|
|
|
|
#endif
|
2022-05-31 07:58:26 -03:00
|
|
|
|
2024-04-04 12:03:27 -03:00
|
|
|
static inline _Py_BackoffCounter
|
2023-12-04 06:42:58 -04:00
|
|
|
adaptive_counter_bits(uint16_t value, uint16_t backoff) {
|
2024-04-04 12:03:27 -03:00
|
|
|
return make_backoff_counter(value, backoff);
|
2022-05-31 07:58:26 -03:00
|
|
|
}
|
|
|
|
|
2024-04-04 12:03:27 -03:00
|
|
|
static inline _Py_BackoffCounter
|
2022-11-09 14:50:09 -04:00
|
|
|
adaptive_counter_warmup(void) {
|
|
|
|
return adaptive_counter_bits(ADAPTIVE_WARMUP_VALUE,
|
|
|
|
ADAPTIVE_WARMUP_BACKOFF);
|
|
|
|
}
|
|
|
|
|
2024-04-04 12:03:27 -03:00
|
|
|
static inline _Py_BackoffCounter
|
2022-11-09 14:50:09 -04:00
|
|
|
adaptive_counter_cooldown(void) {
|
|
|
|
return adaptive_counter_bits(ADAPTIVE_COOLDOWN_VALUE,
|
|
|
|
ADAPTIVE_COOLDOWN_BACKOFF);
|
2022-05-31 07:58:26 -03:00
|
|
|
}
|
|
|
|
|
2024-04-04 12:03:27 -03:00
|
|
|
static inline _Py_BackoffCounter
|
|
|
|
adaptive_counter_backoff(_Py_BackoffCounter counter) {
|
|
|
|
return restart_backoff_counter(counter);
|
2022-05-31 07:58:26 -03:00
|
|
|
}
|
|
|
|
|
2022-06-20 09:00:42 -03:00
|
|
|
|
2023-01-11 16:40:43 -04:00
|
|
|
/* Comparison bit masks. */
|
|
|
|
|
|
|
|
/* Note this evaluates its arguments twice each */
|
|
|
|
#define COMPARISON_BIT(x, y) (1 << (2 * ((x) >= (y)) + ((x) <= (y))))
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The following bits are chosen so that the value of
|
|
|
|
* COMPARSION_BIT(left, right)
|
|
|
|
* masked by the values below will be non-zero if the
|
|
|
|
* comparison is true, and zero if it is false */
|
|
|
|
|
|
|
|
/* This is for values that are unordered, ie. NaN, not types that are unordered, e.g. sets */
|
|
|
|
#define COMPARISON_UNORDERED 1
|
|
|
|
|
|
|
|
#define COMPARISON_LESS_THAN 2
|
|
|
|
#define COMPARISON_GREATER_THAN 4
|
|
|
|
#define COMPARISON_EQUALS 8
|
|
|
|
|
|
|
|
#define COMPARISON_NOT_EQUALS (COMPARISON_UNORDERED | COMPARISON_LESS_THAN | COMPARISON_GREATER_THAN)
|
|
|
|
|
2023-04-12 08:04:55 -03:00
|
|
|
extern int _Py_Instrument(PyCodeObject *co, PyInterpreterState *interp);
|
|
|
|
|
2024-08-13 10:22:57 -03:00
|
|
|
extern _Py_CODEUNIT _Py_GetBaseCodeUnit(PyCodeObject *code, int offset);
|
2023-04-12 08:04:55 -03:00
|
|
|
|
2023-06-02 07:46:18 -03:00
|
|
|
extern int _PyInstruction_GetLength(PyCodeObject *code, int offset);
|
2023-01-11 16:40:43 -04:00
|
|
|
|
2024-08-20 12:52:58 -03:00
|
|
|
struct _PyCode8 _PyCode_DEF(8);
|
|
|
|
|
|
|
|
PyAPI_DATA(const struct _PyCode8) _Py_InitCleanup;
|
|
|
|
|
2024-11-04 15:13:32 -04:00
|
|
|
#ifdef Py_GIL_DISABLED
|
|
|
|
|
|
|
|
// Return a pointer to the thread-local bytecode for the current thread, if it
|
|
|
|
// exists.
|
|
|
|
static inline _Py_CODEUNIT *
|
|
|
|
_PyCode_GetTLBCFast(PyThreadState *tstate, PyCodeObject *co)
|
|
|
|
{
|
|
|
|
_PyCodeArray *code = _Py_atomic_load_ptr_acquire(&co->co_tlbc);
|
|
|
|
int32_t idx = ((_PyThreadStateImpl*) tstate)->tlbc_index;
|
|
|
|
if (idx < code->size && code->entries[idx] != NULL) {
|
|
|
|
return (_Py_CODEUNIT *) code->entries[idx];
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Return a pointer to the thread-local bytecode for the current thread,
|
|
|
|
// creating it if necessary.
|
|
|
|
extern _Py_CODEUNIT *_PyCode_GetTLBC(PyCodeObject *co);
|
|
|
|
|
|
|
|
// Reserve an index for the current thread into thread-local bytecode
|
|
|
|
// arrays
|
|
|
|
//
|
|
|
|
// Returns the reserved index or -1 on error.
|
|
|
|
extern int32_t _Py_ReserveTLBCIndex(PyInterpreterState *interp);
|
|
|
|
|
|
|
|
// Release the current thread's index into thread-local bytecode arrays
|
|
|
|
extern void _Py_ClearTLBCIndex(_PyThreadStateImpl *tstate);
|
|
|
|
|
|
|
|
// Free all TLBC copies not associated with live threads.
|
|
|
|
//
|
|
|
|
// Returns 0 on success or -1 on error.
|
|
|
|
extern int _Py_ClearUnusedTLBC(PyInterpreterState *interp);
|
|
|
|
#endif
|
|
|
|
|
2019-06-03 09:30:58 -03:00
|
|
|
#ifdef __cplusplus
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
#endif /* !Py_INTERNAL_CODE_H */
|