mirror of https://github.com/python/cpython
876 lines
32 KiB
Python
876 lines
32 KiB
Python
import ast
|
|
import os.path
|
|
import re
|
|
from dataclasses import dataclass, field
|
|
from enum import Enum
|
|
from typing import IO, Any, Dict, List, Optional, Set, Text, Tuple
|
|
|
|
from pegen import grammar
|
|
from pegen.grammar import (
|
|
Alt,
|
|
Cut,
|
|
Forced,
|
|
Gather,
|
|
GrammarVisitor,
|
|
Group,
|
|
Leaf,
|
|
Lookahead,
|
|
NamedItem,
|
|
NameLeaf,
|
|
NegativeLookahead,
|
|
Opt,
|
|
PositiveLookahead,
|
|
Repeat0,
|
|
Repeat1,
|
|
Rhs,
|
|
Rule,
|
|
StringLeaf,
|
|
)
|
|
from pegen.parser_generator import ParserGenerator
|
|
|
|
EXTENSION_PREFIX = """\
|
|
#include "pegen.h"
|
|
|
|
#if defined(Py_DEBUG) && defined(Py_BUILD_CORE)
|
|
# define D(x) if (p->debug) { x; }
|
|
#else
|
|
# define D(x)
|
|
#endif
|
|
|
|
#ifdef __wasi__
|
|
# ifdef Py_DEBUG
|
|
# define MAXSTACK 1000
|
|
# else
|
|
# define MAXSTACK 4000
|
|
# endif
|
|
#else
|
|
# define MAXSTACK 6000
|
|
#endif
|
|
|
|
"""
|
|
|
|
|
|
EXTENSION_SUFFIX = """
|
|
void *
|
|
_PyPegen_parse(Parser *p)
|
|
{
|
|
// Initialize keywords
|
|
p->keywords = reserved_keywords;
|
|
p->n_keyword_lists = n_keyword_lists;
|
|
p->soft_keywords = soft_keywords;
|
|
|
|
return start_rule(p);
|
|
}
|
|
"""
|
|
|
|
|
|
class NodeTypes(Enum):
|
|
NAME_TOKEN = 0
|
|
NUMBER_TOKEN = 1
|
|
STRING_TOKEN = 2
|
|
GENERIC_TOKEN = 3
|
|
KEYWORD = 4
|
|
SOFT_KEYWORD = 5
|
|
CUT_OPERATOR = 6
|
|
F_STRING_CHUNK = 7
|
|
|
|
|
|
BASE_NODETYPES = {
|
|
"NAME": NodeTypes.NAME_TOKEN,
|
|
"NUMBER": NodeTypes.NUMBER_TOKEN,
|
|
"STRING": NodeTypes.STRING_TOKEN,
|
|
"SOFT_KEYWORD": NodeTypes.SOFT_KEYWORD,
|
|
}
|
|
|
|
|
|
@dataclass
|
|
class FunctionCall:
|
|
function: str
|
|
arguments: List[Any] = field(default_factory=list)
|
|
assigned_variable: Optional[str] = None
|
|
assigned_variable_type: Optional[str] = None
|
|
return_type: Optional[str] = None
|
|
nodetype: Optional[NodeTypes] = None
|
|
force_true: bool = False
|
|
comment: Optional[str] = None
|
|
|
|
def __str__(self) -> str:
|
|
parts = []
|
|
parts.append(self.function)
|
|
if self.arguments:
|
|
parts.append(f"({', '.join(map(str, self.arguments))})")
|
|
if self.force_true:
|
|
parts.append(", !p->error_indicator")
|
|
if self.assigned_variable:
|
|
if self.assigned_variable_type:
|
|
parts = [
|
|
"(",
|
|
self.assigned_variable,
|
|
" = ",
|
|
"(",
|
|
self.assigned_variable_type,
|
|
")",
|
|
*parts,
|
|
")",
|
|
]
|
|
else:
|
|
parts = ["(", self.assigned_variable, " = ", *parts, ")"]
|
|
if self.comment:
|
|
parts.append(f" // {self.comment}")
|
|
return "".join(parts)
|
|
|
|
|
|
class CCallMakerVisitor(GrammarVisitor):
|
|
def __init__(
|
|
self,
|
|
parser_generator: ParserGenerator,
|
|
exact_tokens: Dict[str, int],
|
|
non_exact_tokens: Set[str],
|
|
):
|
|
self.gen = parser_generator
|
|
self.exact_tokens = exact_tokens
|
|
self.non_exact_tokens = non_exact_tokens
|
|
self.cache: Dict[Any, FunctionCall] = {}
|
|
self.cleanup_statements: List[str] = []
|
|
|
|
def keyword_helper(self, keyword: str) -> FunctionCall:
|
|
return FunctionCall(
|
|
assigned_variable="_keyword",
|
|
function="_PyPegen_expect_token",
|
|
arguments=["p", self.gen.keywords[keyword]],
|
|
return_type="Token *",
|
|
nodetype=NodeTypes.KEYWORD,
|
|
comment=f"token='{keyword}'",
|
|
)
|
|
|
|
def soft_keyword_helper(self, value: str) -> FunctionCall:
|
|
return FunctionCall(
|
|
assigned_variable="_keyword",
|
|
function="_PyPegen_expect_soft_keyword",
|
|
arguments=["p", value],
|
|
return_type="expr_ty",
|
|
nodetype=NodeTypes.SOFT_KEYWORD,
|
|
comment=f"soft_keyword='{value}'",
|
|
)
|
|
|
|
def visit_NameLeaf(self, node: NameLeaf) -> FunctionCall:
|
|
name = node.value
|
|
if name in self.non_exact_tokens:
|
|
if name in BASE_NODETYPES:
|
|
return FunctionCall(
|
|
assigned_variable=f"{name.lower()}_var",
|
|
function=f"_PyPegen_{name.lower()}_token",
|
|
arguments=["p"],
|
|
nodetype=BASE_NODETYPES[name],
|
|
return_type="expr_ty",
|
|
comment=name,
|
|
)
|
|
return FunctionCall(
|
|
assigned_variable=f"{name.lower()}_var",
|
|
function=f"_PyPegen_expect_token",
|
|
arguments=["p", name],
|
|
nodetype=NodeTypes.GENERIC_TOKEN,
|
|
return_type="Token *",
|
|
comment=f"token='{name}'",
|
|
)
|
|
|
|
type = None
|
|
rule = self.gen.all_rules.get(name.lower())
|
|
if rule is not None:
|
|
type = "asdl_seq *" if rule.is_loop() or rule.is_gather() else rule.type
|
|
|
|
return FunctionCall(
|
|
assigned_variable=f"{name}_var",
|
|
function=f"{name}_rule",
|
|
arguments=["p"],
|
|
return_type=type,
|
|
comment=f"{node}",
|
|
)
|
|
|
|
def visit_StringLeaf(self, node: StringLeaf) -> FunctionCall:
|
|
val = ast.literal_eval(node.value)
|
|
if re.match(r"[a-zA-Z_]\w*\Z", val): # This is a keyword
|
|
if node.value.endswith("'"):
|
|
return self.keyword_helper(val)
|
|
else:
|
|
return self.soft_keyword_helper(node.value)
|
|
else:
|
|
assert val in self.exact_tokens, f"{node.value} is not a known literal"
|
|
type = self.exact_tokens[val]
|
|
return FunctionCall(
|
|
assigned_variable="_literal",
|
|
function=f"_PyPegen_expect_token",
|
|
arguments=["p", type],
|
|
nodetype=NodeTypes.GENERIC_TOKEN,
|
|
return_type="Token *",
|
|
comment=f"token='{val}'",
|
|
)
|
|
|
|
def visit_Rhs(self, node: Rhs) -> FunctionCall:
|
|
if node in self.cache:
|
|
return self.cache[node]
|
|
if node.can_be_inlined:
|
|
self.cache[node] = self.generate_call(node.alts[0].items[0])
|
|
else:
|
|
name = self.gen.artifical_rule_from_rhs(node)
|
|
self.cache[node] = FunctionCall(
|
|
assigned_variable=f"{name}_var",
|
|
function=f"{name}_rule",
|
|
arguments=["p"],
|
|
comment=f"{node}",
|
|
)
|
|
return self.cache[node]
|
|
|
|
def visit_NamedItem(self, node: NamedItem) -> FunctionCall:
|
|
call = self.generate_call(node.item)
|
|
if node.name:
|
|
call.assigned_variable = node.name
|
|
if node.type:
|
|
call.assigned_variable_type = node.type
|
|
return call
|
|
|
|
def lookahead_call_helper(self, node: Lookahead, positive: int) -> FunctionCall:
|
|
call = self.generate_call(node.node)
|
|
if call.nodetype == NodeTypes.NAME_TOKEN:
|
|
return FunctionCall(
|
|
function=f"_PyPegen_lookahead_with_name",
|
|
arguments=[positive, call.function, *call.arguments],
|
|
return_type="int",
|
|
)
|
|
elif call.nodetype == NodeTypes.SOFT_KEYWORD:
|
|
return FunctionCall(
|
|
function=f"_PyPegen_lookahead_with_string",
|
|
arguments=[positive, call.function, *call.arguments],
|
|
return_type="int",
|
|
)
|
|
elif call.nodetype in {NodeTypes.GENERIC_TOKEN, NodeTypes.KEYWORD}:
|
|
return FunctionCall(
|
|
function=f"_PyPegen_lookahead_with_int",
|
|
arguments=[positive, call.function, *call.arguments],
|
|
return_type="int",
|
|
comment=f"token={node.node}",
|
|
)
|
|
else:
|
|
return FunctionCall(
|
|
function=f"_PyPegen_lookahead",
|
|
arguments=[positive, call.function, *call.arguments],
|
|
return_type="int",
|
|
)
|
|
|
|
def visit_PositiveLookahead(self, node: PositiveLookahead) -> FunctionCall:
|
|
return self.lookahead_call_helper(node, 1)
|
|
|
|
def visit_NegativeLookahead(self, node: NegativeLookahead) -> FunctionCall:
|
|
return self.lookahead_call_helper(node, 0)
|
|
|
|
def visit_Forced(self, node: Forced) -> FunctionCall:
|
|
call = self.generate_call(node.node)
|
|
if isinstance(node.node, Leaf):
|
|
assert isinstance(node.node, Leaf)
|
|
val = ast.literal_eval(node.node.value)
|
|
assert val in self.exact_tokens, f"{node.node.value} is not a known literal"
|
|
type = self.exact_tokens[val]
|
|
return FunctionCall(
|
|
assigned_variable="_literal",
|
|
function=f"_PyPegen_expect_forced_token",
|
|
arguments=["p", type, f'"{val}"'],
|
|
nodetype=NodeTypes.GENERIC_TOKEN,
|
|
return_type="Token *",
|
|
comment=f"forced_token='{val}'",
|
|
)
|
|
if isinstance(node.node, Group):
|
|
call = self.visit(node.node.rhs)
|
|
call.assigned_variable = None
|
|
call.comment = None
|
|
return FunctionCall(
|
|
assigned_variable="_literal",
|
|
function=f"_PyPegen_expect_forced_result",
|
|
arguments=["p", str(call), f'"{node.node.rhs!s}"'],
|
|
return_type="void *",
|
|
comment=f"forced_token=({node.node.rhs!s})",
|
|
)
|
|
else:
|
|
raise NotImplementedError(f"Forced tokens don't work with {node.node} nodes")
|
|
|
|
def visit_Opt(self, node: Opt) -> FunctionCall:
|
|
call = self.generate_call(node.node)
|
|
return FunctionCall(
|
|
assigned_variable="_opt_var",
|
|
function=call.function,
|
|
arguments=call.arguments,
|
|
force_true=True,
|
|
comment=f"{node}",
|
|
)
|
|
|
|
def visit_Repeat0(self, node: Repeat0) -> FunctionCall:
|
|
if node in self.cache:
|
|
return self.cache[node]
|
|
name = self.gen.artificial_rule_from_repeat(node.node, False)
|
|
self.cache[node] = FunctionCall(
|
|
assigned_variable=f"{name}_var",
|
|
function=f"{name}_rule",
|
|
arguments=["p"],
|
|
return_type="asdl_seq *",
|
|
comment=f"{node}",
|
|
)
|
|
return self.cache[node]
|
|
|
|
def visit_Repeat1(self, node: Repeat1) -> FunctionCall:
|
|
if node in self.cache:
|
|
return self.cache[node]
|
|
name = self.gen.artificial_rule_from_repeat(node.node, True)
|
|
self.cache[node] = FunctionCall(
|
|
assigned_variable=f"{name}_var",
|
|
function=f"{name}_rule",
|
|
arguments=["p"],
|
|
return_type="asdl_seq *",
|
|
comment=f"{node}",
|
|
)
|
|
return self.cache[node]
|
|
|
|
def visit_Gather(self, node: Gather) -> FunctionCall:
|
|
if node in self.cache:
|
|
return self.cache[node]
|
|
name = self.gen.artifical_rule_from_gather(node)
|
|
self.cache[node] = FunctionCall(
|
|
assigned_variable=f"{name}_var",
|
|
function=f"{name}_rule",
|
|
arguments=["p"],
|
|
return_type="asdl_seq *",
|
|
comment=f"{node}",
|
|
)
|
|
return self.cache[node]
|
|
|
|
def visit_Group(self, node: Group) -> FunctionCall:
|
|
return self.generate_call(node.rhs)
|
|
|
|
def visit_Cut(self, node: Cut) -> FunctionCall:
|
|
return FunctionCall(
|
|
assigned_variable="_cut_var",
|
|
return_type="int",
|
|
function="1",
|
|
nodetype=NodeTypes.CUT_OPERATOR,
|
|
)
|
|
|
|
def generate_call(self, node: Any) -> FunctionCall:
|
|
return super().visit(node)
|
|
|
|
|
|
class CParserGenerator(ParserGenerator, GrammarVisitor):
|
|
def __init__(
|
|
self,
|
|
grammar: grammar.Grammar,
|
|
tokens: Dict[int, str],
|
|
exact_tokens: Dict[str, int],
|
|
non_exact_tokens: Set[str],
|
|
file: Optional[IO[Text]],
|
|
debug: bool = False,
|
|
skip_actions: bool = False,
|
|
):
|
|
super().__init__(grammar, set(tokens.values()), file)
|
|
self.callmakervisitor: CCallMakerVisitor = CCallMakerVisitor(
|
|
self, exact_tokens, non_exact_tokens
|
|
)
|
|
self._varname_counter = 0
|
|
self.debug = debug
|
|
self.skip_actions = skip_actions
|
|
self.cleanup_statements: List[str] = []
|
|
|
|
def add_level(self) -> None:
|
|
self.print("if (p->level++ == MAXSTACK) {")
|
|
with self.indent():
|
|
self.print("_Pypegen_stack_overflow(p);")
|
|
self.print("}")
|
|
|
|
def remove_level(self) -> None:
|
|
self.print("p->level--;")
|
|
|
|
def add_return(self, ret_val: str) -> None:
|
|
for stmt in self.cleanup_statements:
|
|
self.print(stmt)
|
|
self.remove_level()
|
|
self.print(f"return {ret_val};")
|
|
|
|
def unique_varname(self, name: str = "tmpvar") -> str:
|
|
new_var = name + "_" + str(self._varname_counter)
|
|
self._varname_counter += 1
|
|
return new_var
|
|
|
|
def call_with_errorcheck_return(self, call_text: str, returnval: str) -> None:
|
|
error_var = self.unique_varname()
|
|
self.print(f"int {error_var} = {call_text};")
|
|
self.print(f"if ({error_var}) {{")
|
|
with self.indent():
|
|
self.add_return(returnval)
|
|
self.print("}")
|
|
|
|
def call_with_errorcheck_goto(self, call_text: str, goto_target: str) -> None:
|
|
error_var = self.unique_varname()
|
|
self.print(f"int {error_var} = {call_text};")
|
|
self.print(f"if ({error_var}) {{")
|
|
with self.indent():
|
|
self.print(f"goto {goto_target};")
|
|
self.print(f"}}")
|
|
|
|
def out_of_memory_return(
|
|
self,
|
|
expr: str,
|
|
cleanup_code: Optional[str] = None,
|
|
) -> None:
|
|
self.print(f"if ({expr}) {{")
|
|
with self.indent():
|
|
if cleanup_code is not None:
|
|
self.print(cleanup_code)
|
|
self.print("p->error_indicator = 1;")
|
|
self.print("PyErr_NoMemory();")
|
|
self.add_return("NULL")
|
|
self.print(f"}}")
|
|
|
|
def out_of_memory_goto(self, expr: str, goto_target: str) -> None:
|
|
self.print(f"if ({expr}) {{")
|
|
with self.indent():
|
|
self.print("PyErr_NoMemory();")
|
|
self.print(f"goto {goto_target};")
|
|
self.print(f"}}")
|
|
|
|
def generate(self, filename: str) -> None:
|
|
self.collect_rules()
|
|
basename = os.path.basename(filename)
|
|
self.print(f"// @generated by pegen from {basename}")
|
|
header = self.grammar.metas.get("header", EXTENSION_PREFIX)
|
|
if header:
|
|
self.print(header.rstrip("\n"))
|
|
subheader = self.grammar.metas.get("subheader", "")
|
|
if subheader:
|
|
self.print(subheader)
|
|
self._setup_keywords()
|
|
self._setup_soft_keywords()
|
|
for i, (rulename, rule) in enumerate(self.all_rules.items(), 1000):
|
|
comment = " // Left-recursive" if rule.left_recursive else ""
|
|
self.print(f"#define {rulename}_type {i}{comment}")
|
|
self.print()
|
|
for rulename, rule in self.all_rules.items():
|
|
if rule.is_loop() or rule.is_gather():
|
|
type = "asdl_seq *"
|
|
elif rule.type:
|
|
type = rule.type + " "
|
|
else:
|
|
type = "void *"
|
|
self.print(f"static {type}{rulename}_rule(Parser *p);")
|
|
self.print()
|
|
for rulename, rule in list(self.all_rules.items()):
|
|
self.print()
|
|
if rule.left_recursive:
|
|
self.print("// Left-recursive")
|
|
self.visit(rule)
|
|
if self.skip_actions:
|
|
mode = 0
|
|
else:
|
|
mode = int(self.rules["start"].type == "mod_ty") if "start" in self.rules else 1
|
|
if mode == 1 and self.grammar.metas.get("bytecode"):
|
|
mode += 1
|
|
modulename = self.grammar.metas.get("modulename", "parse")
|
|
trailer = self.grammar.metas.get("trailer", EXTENSION_SUFFIX)
|
|
if trailer:
|
|
self.print(trailer.rstrip("\n") % dict(mode=mode, modulename=modulename))
|
|
|
|
def _group_keywords_by_length(self) -> Dict[int, List[Tuple[str, int]]]:
|
|
groups: Dict[int, List[Tuple[str, int]]] = {}
|
|
for keyword_str, keyword_type in self.keywords.items():
|
|
length = len(keyword_str)
|
|
if length in groups:
|
|
groups[length].append((keyword_str, keyword_type))
|
|
else:
|
|
groups[length] = [(keyword_str, keyword_type)]
|
|
return groups
|
|
|
|
def _setup_keywords(self) -> None:
|
|
n_keyword_lists = (
|
|
len(max(self.keywords.keys(), key=len)) + 1 if len(self.keywords) > 0 else 0
|
|
)
|
|
self.print(f"static const int n_keyword_lists = {n_keyword_lists};")
|
|
groups = self._group_keywords_by_length()
|
|
self.print("static KeywordToken *reserved_keywords[] = {")
|
|
with self.indent():
|
|
num_groups = max(groups) + 1 if groups else 1
|
|
for keywords_length in range(num_groups):
|
|
if keywords_length not in groups.keys():
|
|
self.print("(KeywordToken[]) {{NULL, -1}},")
|
|
else:
|
|
self.print("(KeywordToken[]) {")
|
|
with self.indent():
|
|
for keyword_str, keyword_type in groups[keywords_length]:
|
|
self.print(f'{{"{keyword_str}", {keyword_type}}},')
|
|
self.print("{NULL, -1},")
|
|
self.print("},")
|
|
self.print("};")
|
|
|
|
def _setup_soft_keywords(self) -> None:
|
|
soft_keywords = sorted(self.soft_keywords)
|
|
self.print("static char *soft_keywords[] = {")
|
|
with self.indent():
|
|
for keyword in soft_keywords:
|
|
self.print(f'"{keyword}",')
|
|
self.print("NULL,")
|
|
self.print("};")
|
|
|
|
def _set_up_token_start_metadata_extraction(self) -> None:
|
|
self.print("if (p->mark == p->fill && _PyPegen_fill_token(p) < 0) {")
|
|
with self.indent():
|
|
self.print("p->error_indicator = 1;")
|
|
self.add_return("NULL")
|
|
self.print("}")
|
|
self.print("int _start_lineno = p->tokens[_mark]->lineno;")
|
|
self.print("UNUSED(_start_lineno); // Only used by EXTRA macro")
|
|
self.print("int _start_col_offset = p->tokens[_mark]->col_offset;")
|
|
self.print("UNUSED(_start_col_offset); // Only used by EXTRA macro")
|
|
|
|
def _set_up_token_end_metadata_extraction(self) -> None:
|
|
self.print("Token *_token = _PyPegen_get_last_nonnwhitespace_token(p);")
|
|
self.print("if (_token == NULL) {")
|
|
with self.indent():
|
|
self.add_return("NULL")
|
|
self.print("}")
|
|
self.print("int _end_lineno = _token->end_lineno;")
|
|
self.print("UNUSED(_end_lineno); // Only used by EXTRA macro")
|
|
self.print("int _end_col_offset = _token->end_col_offset;")
|
|
self.print("UNUSED(_end_col_offset); // Only used by EXTRA macro")
|
|
|
|
def _check_for_errors(self) -> None:
|
|
self.print("if (p->error_indicator) {")
|
|
with self.indent():
|
|
self.add_return("NULL")
|
|
self.print("}")
|
|
|
|
def _set_up_rule_memoization(self, node: Rule, result_type: str) -> None:
|
|
self.print("{")
|
|
with self.indent():
|
|
self.add_level()
|
|
self.print(f"{result_type} _res = NULL;")
|
|
self.print(f"if (_PyPegen_is_memoized(p, {node.name}_type, &_res)) {{")
|
|
with self.indent():
|
|
self.add_return("_res")
|
|
self.print("}")
|
|
self.print("int _mark = p->mark;")
|
|
self.print("int _resmark = p->mark;")
|
|
self.print("while (1) {")
|
|
with self.indent():
|
|
self.call_with_errorcheck_return(
|
|
f"_PyPegen_update_memo(p, _mark, {node.name}_type, _res)", "_res"
|
|
)
|
|
self.print("p->mark = _mark;")
|
|
self.print(f"void *_raw = {node.name}_raw(p);")
|
|
self.print("if (p->error_indicator) {")
|
|
with self.indent():
|
|
self.add_return("NULL")
|
|
self.print("}")
|
|
self.print("if (_raw == NULL || p->mark <= _resmark)")
|
|
with self.indent():
|
|
self.print("break;")
|
|
self.print(f"_resmark = p->mark;")
|
|
self.print("_res = _raw;")
|
|
self.print("}")
|
|
self.print(f"p->mark = _resmark;")
|
|
self.add_return("_res")
|
|
self.print("}")
|
|
self.print(f"static {result_type}")
|
|
self.print(f"{node.name}_raw(Parser *p)")
|
|
|
|
def _should_memoize(self, node: Rule) -> bool:
|
|
return node.memo and not node.left_recursive
|
|
|
|
def _handle_default_rule_body(self, node: Rule, rhs: Rhs, result_type: str) -> None:
|
|
memoize = self._should_memoize(node)
|
|
|
|
with self.indent():
|
|
self.add_level()
|
|
self._check_for_errors()
|
|
self.print(f"{result_type} _res = NULL;")
|
|
if memoize:
|
|
self.print(f"if (_PyPegen_is_memoized(p, {node.name}_type, &_res)) {{")
|
|
with self.indent():
|
|
self.add_return("_res")
|
|
self.print("}")
|
|
self.print("int _mark = p->mark;")
|
|
if any(alt.action and "EXTRA" in alt.action for alt in rhs.alts):
|
|
self._set_up_token_start_metadata_extraction()
|
|
self.visit(
|
|
rhs,
|
|
is_loop=False,
|
|
is_gather=node.is_gather(),
|
|
rulename=node.name,
|
|
)
|
|
if self.debug:
|
|
self.print(f'D(fprintf(stderr, "Fail at %d: {node.name}\\n", p->mark));')
|
|
self.print("_res = NULL;")
|
|
self.print(" done:")
|
|
with self.indent():
|
|
if memoize:
|
|
self.print(f"_PyPegen_insert_memo(p, _mark, {node.name}_type, _res);")
|
|
self.add_return("_res")
|
|
|
|
def _handle_loop_rule_body(self, node: Rule, rhs: Rhs) -> None:
|
|
memoize = self._should_memoize(node)
|
|
is_repeat1 = node.name.startswith("_loop1")
|
|
|
|
with self.indent():
|
|
self.add_level()
|
|
self._check_for_errors()
|
|
self.print("void *_res = NULL;")
|
|
if memoize:
|
|
self.print(f"if (_PyPegen_is_memoized(p, {node.name}_type, &_res)) {{")
|
|
with self.indent():
|
|
self.add_return("_res")
|
|
self.print("}")
|
|
self.print("int _mark = p->mark;")
|
|
if memoize:
|
|
self.print("int _start_mark = p->mark;")
|
|
self.print("void **_children = PyMem_Malloc(sizeof(void *));")
|
|
self.out_of_memory_return(f"!_children")
|
|
self.print("Py_ssize_t _children_capacity = 1;")
|
|
self.print("Py_ssize_t _n = 0;")
|
|
if any(alt.action and "EXTRA" in alt.action for alt in rhs.alts):
|
|
self._set_up_token_start_metadata_extraction()
|
|
self.visit(
|
|
rhs,
|
|
is_loop=True,
|
|
is_gather=node.is_gather(),
|
|
rulename=node.name,
|
|
)
|
|
if is_repeat1:
|
|
self.print("if (_n == 0 || p->error_indicator) {")
|
|
with self.indent():
|
|
self.print("PyMem_Free(_children);")
|
|
self.add_return("NULL")
|
|
self.print("}")
|
|
self.print("asdl_seq *_seq = (asdl_seq*)_Py_asdl_generic_seq_new(_n, p->arena);")
|
|
self.out_of_memory_return(f"!_seq", cleanup_code="PyMem_Free(_children);")
|
|
self.print("for (int i = 0; i < _n; i++) asdl_seq_SET_UNTYPED(_seq, i, _children[i]);")
|
|
self.print("PyMem_Free(_children);")
|
|
if memoize and node.name:
|
|
self.print(f"_PyPegen_insert_memo(p, _start_mark, {node.name}_type, _seq);")
|
|
self.add_return("_seq")
|
|
|
|
def visit_Rule(self, node: Rule) -> None:
|
|
is_loop = node.is_loop()
|
|
is_gather = node.is_gather()
|
|
rhs = node.flatten()
|
|
if is_loop or is_gather:
|
|
result_type = "asdl_seq *"
|
|
elif node.type:
|
|
result_type = node.type
|
|
else:
|
|
result_type = "void *"
|
|
|
|
for line in str(node).splitlines():
|
|
self.print(f"// {line}")
|
|
if node.left_recursive and node.leader:
|
|
self.print(f"static {result_type} {node.name}_raw(Parser *);")
|
|
|
|
self.print(f"static {result_type}")
|
|
self.print(f"{node.name}_rule(Parser *p)")
|
|
|
|
if node.left_recursive and node.leader:
|
|
self._set_up_rule_memoization(node, result_type)
|
|
|
|
self.print("{")
|
|
|
|
if node.name.endswith("without_invalid"):
|
|
with self.indent():
|
|
self.print("int _prev_call_invalid = p->call_invalid_rules;")
|
|
self.print("p->call_invalid_rules = 0;")
|
|
self.cleanup_statements.append("p->call_invalid_rules = _prev_call_invalid;")
|
|
|
|
if is_loop:
|
|
self._handle_loop_rule_body(node, rhs)
|
|
else:
|
|
self._handle_default_rule_body(node, rhs, result_type)
|
|
|
|
if node.name.endswith("without_invalid"):
|
|
self.cleanup_statements.pop()
|
|
|
|
self.print("}")
|
|
|
|
def visit_NamedItem(self, node: NamedItem) -> None:
|
|
call = self.callmakervisitor.generate_call(node)
|
|
if call.assigned_variable:
|
|
call.assigned_variable = self.dedupe(call.assigned_variable)
|
|
self.print(call)
|
|
|
|
def visit_Rhs(
|
|
self, node: Rhs, is_loop: bool, is_gather: bool, rulename: Optional[str]
|
|
) -> None:
|
|
if is_loop:
|
|
assert len(node.alts) == 1
|
|
for alt in node.alts:
|
|
self.visit(alt, is_loop=is_loop, is_gather=is_gather, rulename=rulename)
|
|
|
|
def join_conditions(self, keyword: str, node: Any) -> None:
|
|
self.print(f"{keyword} (")
|
|
with self.indent():
|
|
first = True
|
|
for item in node.items:
|
|
if first:
|
|
first = False
|
|
else:
|
|
self.print("&&")
|
|
self.visit(item)
|
|
self.print(")")
|
|
|
|
def emit_action(self, node: Alt, cleanup_code: Optional[str] = None) -> None:
|
|
self.print(f"_res = {node.action};")
|
|
|
|
self.print("if (_res == NULL && PyErr_Occurred()) {")
|
|
with self.indent():
|
|
self.print("p->error_indicator = 1;")
|
|
if cleanup_code:
|
|
self.print(cleanup_code)
|
|
self.add_return("NULL")
|
|
self.print("}")
|
|
|
|
if self.debug:
|
|
self.print(
|
|
f'D(fprintf(stderr, "Hit with action [%d-%d]: %s\\n", _mark, p->mark, "{node}"));'
|
|
)
|
|
|
|
def emit_default_action(self, is_gather: bool, node: Alt) -> None:
|
|
if len(self.local_variable_names) > 1:
|
|
if is_gather:
|
|
assert len(self.local_variable_names) == 2
|
|
self.print(
|
|
f"_res = _PyPegen_seq_insert_in_front(p, "
|
|
f"{self.local_variable_names[0]}, {self.local_variable_names[1]});"
|
|
)
|
|
else:
|
|
if self.debug:
|
|
self.print(
|
|
f'D(fprintf(stderr, "Hit without action [%d:%d]: %s\\n", _mark, p->mark, "{node}"));'
|
|
)
|
|
self.print(
|
|
f"_res = _PyPegen_dummy_name(p, {', '.join(self.local_variable_names)});"
|
|
)
|
|
else:
|
|
if self.debug:
|
|
self.print(
|
|
f'D(fprintf(stderr, "Hit with default action [%d:%d]: %s\\n", _mark, p->mark, "{node}"));'
|
|
)
|
|
self.print(f"_res = {self.local_variable_names[0]};")
|
|
|
|
def emit_dummy_action(self) -> None:
|
|
self.print("_res = _PyPegen_dummy_name(p);")
|
|
|
|
def handle_alt_normal(self, node: Alt, is_gather: bool, rulename: Optional[str]) -> None:
|
|
self.join_conditions(keyword="if", node=node)
|
|
self.print("{")
|
|
# We have parsed successfully all the conditions for the option.
|
|
with self.indent():
|
|
node_str = str(node).replace('"', '\\"')
|
|
self.print(
|
|
f'D(fprintf(stderr, "%*c+ {rulename}[%d-%d]: %s succeeded!\\n", p->level, \' \', _mark, p->mark, "{node_str}"));'
|
|
)
|
|
# Prepare to emit the rule action and do so
|
|
if node.action and "EXTRA" in node.action:
|
|
self._set_up_token_end_metadata_extraction()
|
|
if self.skip_actions:
|
|
self.emit_dummy_action()
|
|
elif node.action:
|
|
self.emit_action(node)
|
|
else:
|
|
self.emit_default_action(is_gather, node)
|
|
|
|
# As the current option has parsed correctly, do not continue with the rest.
|
|
self.print(f"goto done;")
|
|
self.print("}")
|
|
|
|
def handle_alt_loop(self, node: Alt, is_gather: bool, rulename: Optional[str]) -> None:
|
|
# Condition of the main body of the alternative
|
|
self.join_conditions(keyword="while", node=node)
|
|
self.print("{")
|
|
# We have parsed successfully one item!
|
|
with self.indent():
|
|
# Prepare to emit the rule action and do so
|
|
if node.action and "EXTRA" in node.action:
|
|
self._set_up_token_end_metadata_extraction()
|
|
if self.skip_actions:
|
|
self.emit_dummy_action()
|
|
elif node.action:
|
|
self.emit_action(node, cleanup_code="PyMem_Free(_children);")
|
|
else:
|
|
self.emit_default_action(is_gather, node)
|
|
|
|
# Add the result of rule to the temporary buffer of children. This buffer
|
|
# will populate later an asdl_seq with all elements to return.
|
|
self.print("if (_n == _children_capacity) {")
|
|
with self.indent():
|
|
self.print("_children_capacity *= 2;")
|
|
self.print(
|
|
"void **_new_children = PyMem_Realloc(_children, _children_capacity*sizeof(void *));"
|
|
)
|
|
self.out_of_memory_return(f"!_new_children", cleanup_code="PyMem_Free(_children);")
|
|
self.print("_children = _new_children;")
|
|
self.print("}")
|
|
self.print("_children[_n++] = _res;")
|
|
self.print("_mark = p->mark;")
|
|
self.print("}")
|
|
|
|
def visit_Alt(
|
|
self, node: Alt, is_loop: bool, is_gather: bool, rulename: Optional[str]
|
|
) -> None:
|
|
if len(node.items) == 1 and str(node.items[0]).startswith("invalid_"):
|
|
self.print(f"if (p->call_invalid_rules) {{ // {node}")
|
|
else:
|
|
self.print(f"{{ // {node}")
|
|
with self.indent():
|
|
self._check_for_errors()
|
|
node_str = str(node).replace('"', '\\"')
|
|
self.print(
|
|
f'D(fprintf(stderr, "%*c> {rulename}[%d-%d]: %s\\n", p->level, \' \', _mark, p->mark, "{node_str}"));'
|
|
)
|
|
# Prepare variable declarations for the alternative
|
|
vars = self.collect_vars(node)
|
|
for v, var_type in sorted(item for item in vars.items() if item[0] is not None):
|
|
if not var_type:
|
|
var_type = "void *"
|
|
else:
|
|
var_type += " "
|
|
if v == "_cut_var":
|
|
v += " = 0" # cut_var must be initialized
|
|
self.print(f"{var_type}{v};")
|
|
if v and v.startswith("_opt_var"):
|
|
self.print(f"UNUSED({v}); // Silence compiler warnings")
|
|
|
|
with self.local_variable_context():
|
|
if is_loop:
|
|
self.handle_alt_loop(node, is_gather, rulename)
|
|
else:
|
|
self.handle_alt_normal(node, is_gather, rulename)
|
|
|
|
self.print("p->mark = _mark;")
|
|
node_str = str(node).replace('"', '\\"')
|
|
self.print(
|
|
f"D(fprintf(stderr, \"%*c%s {rulename}[%d-%d]: %s failed!\\n\", p->level, ' ',\n"
|
|
f' p->error_indicator ? "ERROR!" : "-", _mark, p->mark, "{node_str}"));'
|
|
)
|
|
if "_cut_var" in vars:
|
|
self.print("if (_cut_var) {")
|
|
with self.indent():
|
|
self.add_return("NULL")
|
|
self.print("}")
|
|
self.print("}")
|
|
|
|
def collect_vars(self, node: Alt) -> Dict[Optional[str], Optional[str]]:
|
|
types = {}
|
|
with self.local_variable_context():
|
|
for item in node.items:
|
|
name, type = self.add_var(item)
|
|
types[name] = type
|
|
return types
|
|
|
|
def add_var(self, node: NamedItem) -> Tuple[Optional[str], Optional[str]]:
|
|
call = self.callmakervisitor.generate_call(node.item)
|
|
name = node.name if node.name else call.assigned_variable
|
|
if name is not None:
|
|
name = self.dedupe(name)
|
|
return_type = call.return_type if node.type is None else node.type
|
|
return name, return_type
|