mirror of https://github.com/python/cpython
220 lines
5.3 KiB
Python
220 lines
5.3 KiB
Python
from pathlib import Path
|
|
from typing import TextIO
|
|
|
|
from analyzer import (
|
|
Instruction,
|
|
Uop,
|
|
analyze_files,
|
|
Properties,
|
|
Skip,
|
|
)
|
|
from cwriter import CWriter
|
|
from typing import Callable, Mapping, TextIO, Iterator
|
|
from lexer import Token
|
|
from stack import StackOffset, Stack
|
|
|
|
|
|
ROOT = Path(__file__).parent.parent.parent
|
|
DEFAULT_INPUT = (ROOT / "Python/bytecodes.c").absolute().as_posix()
|
|
|
|
|
|
def root_relative_path(filename: str) -> str:
|
|
try:
|
|
return Path(filename).absolute().relative_to(ROOT).as_posix()
|
|
except ValueError:
|
|
# Not relative to root, just return original path.
|
|
return filename
|
|
|
|
|
|
def write_header(
|
|
generator: str, sources: list[str], outfile: TextIO, comment: str = "//"
|
|
) -> None:
|
|
outfile.write(
|
|
f"""{comment} This file is generated by {root_relative_path(generator)}
|
|
{comment} from:
|
|
{comment} {", ".join(root_relative_path(src) for src in sources)}
|
|
{comment} Do not edit!
|
|
"""
|
|
)
|
|
|
|
|
|
def emit_to(out: CWriter, tkn_iter: Iterator[Token], end: str) -> None:
|
|
parens = 0
|
|
for tkn in tkn_iter:
|
|
if tkn.kind == end and parens == 0:
|
|
return
|
|
if tkn.kind == "LPAREN":
|
|
parens += 1
|
|
if tkn.kind == "RPAREN":
|
|
parens -= 1
|
|
out.emit(tkn)
|
|
|
|
|
|
def replace_deopt(
|
|
out: CWriter,
|
|
tkn: Token,
|
|
tkn_iter: Iterator[Token],
|
|
uop: Uop,
|
|
unused: Stack,
|
|
inst: Instruction | None,
|
|
) -> None:
|
|
out.emit_at("DEOPT_IF", tkn)
|
|
out.emit(next(tkn_iter))
|
|
emit_to(out, tkn_iter, "RPAREN")
|
|
next(tkn_iter) # Semi colon
|
|
out.emit(", ")
|
|
assert inst is not None
|
|
assert inst.family is not None
|
|
out.emit(inst.family.name)
|
|
out.emit(");\n")
|
|
|
|
|
|
def replace_error(
|
|
out: CWriter,
|
|
tkn: Token,
|
|
tkn_iter: Iterator[Token],
|
|
uop: Uop,
|
|
stack: Stack,
|
|
inst: Instruction | None,
|
|
) -> None:
|
|
out.emit_at("if ", tkn)
|
|
out.emit(next(tkn_iter))
|
|
emit_to(out, tkn_iter, "COMMA")
|
|
label = next(tkn_iter).text
|
|
next(tkn_iter) # RPAREN
|
|
next(tkn_iter) # Semi colon
|
|
out.emit(") ")
|
|
c_offset = stack.peek_offset.to_c()
|
|
try:
|
|
offset = -int(c_offset)
|
|
close = ";\n"
|
|
except ValueError:
|
|
offset = None
|
|
out.emit(f"{{ stack_pointer += {c_offset}; ")
|
|
close = "; }\n"
|
|
out.emit("goto ")
|
|
if offset:
|
|
out.emit(f"pop_{offset}_")
|
|
out.emit(label)
|
|
out.emit(close)
|
|
|
|
|
|
def replace_decrefs(
|
|
out: CWriter,
|
|
tkn: Token,
|
|
tkn_iter: Iterator[Token],
|
|
uop: Uop,
|
|
stack: Stack,
|
|
inst: Instruction | None,
|
|
) -> None:
|
|
next(tkn_iter)
|
|
next(tkn_iter)
|
|
next(tkn_iter)
|
|
out.emit_at("", tkn)
|
|
for var in uop.stack.inputs:
|
|
if var.name == "unused" or var.name == "null" or var.peek:
|
|
continue
|
|
if var.size != "1":
|
|
out.emit(f"for (int _i = {var.size}; --_i >= 0;) {{\n")
|
|
out.emit(f"Py_DECREF({var.name}[_i]);\n")
|
|
out.emit("}\n")
|
|
elif var.condition:
|
|
out.emit(f"Py_XDECREF({var.name});\n")
|
|
else:
|
|
out.emit(f"Py_DECREF({var.name});\n")
|
|
|
|
|
|
def replace_sync_sp(
|
|
out: CWriter,
|
|
tkn: Token,
|
|
tkn_iter: Iterator[Token],
|
|
uop: Uop,
|
|
stack: Stack,
|
|
inst: Instruction | None,
|
|
) -> None:
|
|
next(tkn_iter)
|
|
next(tkn_iter)
|
|
next(tkn_iter)
|
|
stack.flush(out)
|
|
|
|
|
|
def replace_check_eval_breaker(
|
|
out: CWriter,
|
|
tkn: Token,
|
|
tkn_iter: Iterator[Token],
|
|
uop: Uop,
|
|
stack: Stack,
|
|
inst: Instruction | None,
|
|
) -> None:
|
|
next(tkn_iter)
|
|
next(tkn_iter)
|
|
next(tkn_iter)
|
|
if not uop.properties.ends_with_eval_breaker:
|
|
out.emit_at("CHECK_EVAL_BREAKER();", tkn)
|
|
|
|
|
|
REPLACEMENT_FUNCTIONS = {
|
|
"DEOPT_IF": replace_deopt,
|
|
"ERROR_IF": replace_error,
|
|
"DECREF_INPUTS": replace_decrefs,
|
|
"CHECK_EVAL_BREAKER": replace_check_eval_breaker,
|
|
"SYNC_SP": replace_sync_sp,
|
|
}
|
|
|
|
ReplacementFunctionType = Callable[
|
|
[CWriter, Token, Iterator[Token], Uop, Stack, Instruction | None], None
|
|
]
|
|
|
|
|
|
def emit_tokens(
|
|
out: CWriter,
|
|
uop: Uop,
|
|
stack: Stack,
|
|
inst: Instruction | None,
|
|
replacement_functions: Mapping[
|
|
str, ReplacementFunctionType
|
|
] = REPLACEMENT_FUNCTIONS,
|
|
) -> None:
|
|
tkns = uop.body[1:-1]
|
|
if not tkns:
|
|
return
|
|
tkn_iter = iter(tkns)
|
|
out.start_line()
|
|
for tkn in tkn_iter:
|
|
if tkn.kind == "IDENTIFIER" and tkn.text in replacement_functions:
|
|
replacement_functions[tkn.text](out, tkn, tkn_iter, uop, stack, inst)
|
|
else:
|
|
out.emit(tkn)
|
|
|
|
|
|
def cflags(p: Properties) -> str:
|
|
flags: list[str] = []
|
|
if p.oparg:
|
|
flags.append("HAS_ARG_FLAG")
|
|
if p.uses_co_consts:
|
|
flags.append("HAS_CONST_FLAG")
|
|
if p.uses_co_names:
|
|
flags.append("HAS_NAME_FLAG")
|
|
if p.jumps:
|
|
flags.append("HAS_JUMP_FLAG")
|
|
if p.has_free:
|
|
flags.append("HAS_FREE_FLAG")
|
|
if p.uses_locals:
|
|
flags.append("HAS_LOCAL_FLAG")
|
|
if p.eval_breaker:
|
|
flags.append("HAS_EVAL_BREAK_FLAG")
|
|
if p.deopts:
|
|
flags.append("HAS_DEOPT_FLAG")
|
|
if not p.infallible:
|
|
flags.append("HAS_ERROR_FLAG")
|
|
if p.escapes:
|
|
flags.append("HAS_ESCAPES_FLAG")
|
|
if p.pure:
|
|
flags.append("HAS_PURE_FLAG")
|
|
if p.passthrough:
|
|
flags.append("HAS_PASSTHROUGH_FLAG")
|
|
if flags:
|
|
return " | ".join(flags)
|
|
else:
|
|
return "0"
|