bpo-30406: Make async and await proper keywords (#1669)

Per PEP 492, 'async' and 'await' should become proper keywords in 3.7.
This commit is contained in:
Jelle Zijlstra 2017-10-05 20:24:46 -07:00 committed by Yury Selivanov
parent 2084b30e54
commit ac317700ce
22 changed files with 419 additions and 623 deletions

View File

@ -98,8 +98,6 @@ The token constants are:
RARROW RARROW
ELLIPSIS ELLIPSIS
OP OP
AWAIT
ASYNC
ERRORTOKEN ERRORTOKEN
N_TOKENS N_TOKENS
NT_OFFSET NT_OFFSET
@ -129,9 +127,11 @@ the :mod:`tokenize` module.
.. versionchanged:: 3.5 .. versionchanged:: 3.5
Added :data:`AWAIT` and :data:`ASYNC` tokens. Starting with Added :data:`AWAIT` and :data:`ASYNC` tokens.
Python 3.7, "async" and "await" will be tokenized as :data:`NAME`
tokens, and :data:`AWAIT` and :data:`ASYNC` will be removed.
.. versionchanged:: 3.7 .. versionchanged:: 3.7
Added :data:`COMMENT`, :data:`NL` and :data:`ENCODING` tokens. Added :data:`COMMENT`, :data:`NL` and :data:`ENCODING` tokens.
.. versionchanged:: 3.7
Removed :data:`AWAIT` and :data:`ASYNC` tokens. "async" and "await" are
now tokenized as :data:`NAME` tokens.

View File

@ -272,9 +272,9 @@ class MiscNews(Directive):
# Support for building "topic help" for pydoc # Support for building "topic help" for pydoc
pydoc_topic_labels = [ pydoc_topic_labels = [
'assert', 'assignment', 'atom-identifiers', 'atom-literals', 'assert', 'assignment', 'async', 'atom-identifiers', 'atom-literals',
'attribute-access', 'attribute-references', 'augassign', 'binary', 'attribute-access', 'attribute-references', 'augassign', 'await',
'bitwise', 'bltin-code-objects', 'bltin-ellipsis-object', 'binary', 'bitwise', 'bltin-code-objects', 'bltin-ellipsis-object',
'bltin-null-object', 'bltin-type-objects', 'booleans', 'bltin-null-object', 'bltin-type-objects', 'booleans',
'break', 'callable-types', 'calls', 'class', 'comparisons', 'compound', 'break', 'callable-types', 'calls', 'class', 'comparisons', 'compound',
'context-managers', 'continue', 'conversions', 'customization', 'debugger', 'context-managers', 'continue', 'conversions', 'customization', 'debugger',

View File

@ -16,7 +16,7 @@ decorator: '@' dotted_name [ '(' [arglist] ')' ] NEWLINE
decorators: decorator+ decorators: decorator+
decorated: decorators (classdef | funcdef | async_funcdef) decorated: decorators (classdef | funcdef | async_funcdef)
async_funcdef: ASYNC funcdef async_funcdef: 'async' funcdef
funcdef: 'def' NAME parameters ['->' test] ':' suite funcdef: 'def' NAME parameters ['->' test] ':' suite
parameters: '(' [typedargslist] ')' parameters: '(' [typedargslist] ')'
@ -68,7 +68,7 @@ nonlocal_stmt: 'nonlocal' NAME (',' NAME)*
assert_stmt: 'assert' test [',' test] assert_stmt: 'assert' test [',' test]
compound_stmt: if_stmt | while_stmt | for_stmt | try_stmt | with_stmt | funcdef | classdef | decorated | async_stmt compound_stmt: if_stmt | while_stmt | for_stmt | try_stmt | with_stmt | funcdef | classdef | decorated | async_stmt
async_stmt: ASYNC (funcdef | with_stmt | for_stmt) async_stmt: 'async' (funcdef | with_stmt | for_stmt)
if_stmt: 'if' test ':' suite ('elif' test ':' suite)* ['else' ':' suite] if_stmt: 'if' test ':' suite ('elif' test ':' suite)* ['else' ':' suite]
while_stmt: 'while' test ':' suite ['else' ':' suite] while_stmt: 'while' test ':' suite ['else' ':' suite]
for_stmt: 'for' exprlist 'in' testlist ':' suite ['else' ':' suite] for_stmt: 'for' exprlist 'in' testlist ':' suite ['else' ':' suite]
@ -103,7 +103,7 @@ arith_expr: term (('+'|'-') term)*
term: factor (('*'|'@'|'/'|'%'|'//') factor)* term: factor (('*'|'@'|'/'|'%'|'//') factor)*
factor: ('+'|'-'|'~') factor | power factor: ('+'|'-'|'~') factor | power
power: atom_expr ['**' factor] power: atom_expr ['**' factor]
atom_expr: [AWAIT] atom trailer* atom_expr: ['await'] atom trailer*
atom: ('(' [yield_expr|testlist_comp] ')' | atom: ('(' [yield_expr|testlist_comp] ')' |
'[' [testlist_comp] ']' | '[' [testlist_comp] ']' |
'{' [dictorsetmaker] '}' | '{' [dictorsetmaker] '}' |
@ -139,7 +139,8 @@ argument: ( test [comp_for] |
'*' test ) '*' test )
comp_iter: comp_for | comp_if comp_iter: comp_for | comp_if
comp_for: [ASYNC] 'for' exprlist 'in' or_test [comp_iter] sync_comp_for: 'for' exprlist 'in' or_test [comp_iter]
comp_for: ['async'] sync_comp_for
comp_if: 'if' test_nocond [comp_iter] comp_if: 'if' test_nocond [comp_iter]
# not used in grammar, but may appear in "node" passed from Parser to Compiler # not used in grammar, but may appear in "node" passed from Parser to Compiler

View File

@ -81,8 +81,9 @@
#define arglist 334 #define arglist 334
#define argument 335 #define argument 335
#define comp_iter 336 #define comp_iter 336
#define comp_for 337 #define sync_comp_for 337
#define comp_if 338 #define comp_for 338
#define encoding_decl 339 #define comp_if 339
#define yield_expr 340 #define encoding_decl 340
#define yield_arg 341 #define yield_expr 341
#define yield_arg 342

View File

@ -9,77 +9,75 @@ extern "C" {
#undef TILDE /* Prevent clash of our definition with system macro. Ex AIX, ioctl.h */ #undef TILDE /* Prevent clash of our definition with system macro. Ex AIX, ioctl.h */
#define ENDMARKER 0 #define ENDMARKER 0
#define NAME 1 #define NAME 1
#define NUMBER 2 #define NUMBER 2
#define STRING 3 #define STRING 3
#define NEWLINE 4 #define NEWLINE 4
#define INDENT 5 #define INDENT 5
#define DEDENT 6 #define DEDENT 6
#define LPAR 7 #define LPAR 7
#define RPAR 8 #define RPAR 8
#define LSQB 9 #define LSQB 9
#define RSQB 10 #define RSQB 10
#define COLON 11 #define COLON 11
#define COMMA 12 #define COMMA 12
#define SEMI 13 #define SEMI 13
#define PLUS 14 #define PLUS 14
#define MINUS 15 #define MINUS 15
#define STAR 16 #define STAR 16
#define SLASH 17 #define SLASH 17
#define VBAR 18 #define VBAR 18
#define AMPER 19 #define AMPER 19
#define LESS 20 #define LESS 20
#define GREATER 21 #define GREATER 21
#define EQUAL 22 #define EQUAL 22
#define DOT 23 #define DOT 23
#define PERCENT 24 #define PERCENT 24
#define LBRACE 25 #define LBRACE 25
#define RBRACE 26 #define RBRACE 26
#define EQEQUAL 27 #define EQEQUAL 27
#define NOTEQUAL 28 #define NOTEQUAL 28
#define LESSEQUAL 29 #define LESSEQUAL 29
#define GREATEREQUAL 30 #define GREATEREQUAL 30
#define TILDE 31 #define TILDE 31
#define CIRCUMFLEX 32 #define CIRCUMFLEX 32
#define LEFTSHIFT 33 #define LEFTSHIFT 33
#define RIGHTSHIFT 34 #define RIGHTSHIFT 34
#define DOUBLESTAR 35 #define DOUBLESTAR 35
#define PLUSEQUAL 36 #define PLUSEQUAL 36
#define MINEQUAL 37 #define MINEQUAL 37
#define STAREQUAL 38 #define STAREQUAL 38
#define SLASHEQUAL 39 #define SLASHEQUAL 39
#define PERCENTEQUAL 40 #define PERCENTEQUAL 40
#define AMPEREQUAL 41 #define AMPEREQUAL 41
#define VBAREQUAL 42 #define VBAREQUAL 42
#define CIRCUMFLEXEQUAL 43 #define CIRCUMFLEXEQUAL 43
#define LEFTSHIFTEQUAL 44 #define LEFTSHIFTEQUAL 44
#define RIGHTSHIFTEQUAL 45 #define RIGHTSHIFTEQUAL 45
#define DOUBLESTAREQUAL 46 #define DOUBLESTAREQUAL 46
#define DOUBLESLASH 47 #define DOUBLESLASH 47
#define DOUBLESLASHEQUAL 48 #define DOUBLESLASHEQUAL 48
#define AT 49 #define AT 49
#define ATEQUAL 50 #define ATEQUAL 50
#define RARROW 51 #define RARROW 51
#define ELLIPSIS 52 #define ELLIPSIS 52
/* Don't forget to update the table _PyParser_TokenNames in tokenizer.c! */ /* Don't forget to update the table _PyParser_TokenNames in tokenizer.c! */
#define OP 53 #define OP 53
#define AWAIT 54 #define ERRORTOKEN 54
#define ASYNC 55
#define ERRORTOKEN 56
/* These aren't used by the C tokenizer but are needed for tokenize.py */ /* These aren't used by the C tokenizer but are needed for tokenize.py */
#define COMMENT 57 #define COMMENT 55
#define NL 58 #define NL 56
#define ENCODING 59 #define ENCODING 57
#define N_TOKENS 60 #define N_TOKENS 58
/* Special definitions for cooperation with parser */ /* Special definitions for cooperation with parser */
#define NT_OFFSET 256 #define NT_OFFSET 256
#define ISTERMINAL(x) ((x) < NT_OFFSET) #define ISTERMINAL(x) ((x) < NT_OFFSET)
#define ISNONTERMINAL(x) ((x) >= NT_OFFSET) #define ISNONTERMINAL(x) ((x) >= NT_OFFSET)
#define ISEOF(x) ((x) == ENDMARKER) #define ISEOF(x) ((x) == ENDMARKER)
PyAPI_DATA(const char *) _PyParser_TokenNames[]; /* Token names */ PyAPI_DATA(const char *) _PyParser_TokenNames[]; /* Token names */

View File

@ -20,6 +20,8 @@ kwlist = [
'and', 'and',
'as', 'as',
'assert', 'assert',
'async',
'await',
'break', 'break',
'class', 'class',
'continue', 'continue',

View File

@ -34,7 +34,7 @@ eval_input: testlist NEWLINE* ENDMARKER
decorator: '@' dotted_name [ '(' [arglist] ')' ] NEWLINE decorator: '@' dotted_name [ '(' [arglist] ')' ] NEWLINE
decorators: decorator+ decorators: decorator+
decorated: decorators (classdef | funcdef | async_funcdef) decorated: decorators (classdef | funcdef | async_funcdef)
async_funcdef: ASYNC funcdef async_funcdef: 'async' funcdef
funcdef: 'def' NAME parameters ['->' test] ':' suite funcdef: 'def' NAME parameters ['->' test] ':' suite
parameters: '(' [typedargslist] ')' parameters: '(' [typedargslist] ')'
typedargslist: ((tfpdef ['=' test] ',')* typedargslist: ((tfpdef ['=' test] ',')*
@ -85,7 +85,7 @@ exec_stmt: 'exec' expr ['in' test [',' test]]
assert_stmt: 'assert' test [',' test] assert_stmt: 'assert' test [',' test]
compound_stmt: if_stmt | while_stmt | for_stmt | try_stmt | with_stmt | funcdef | classdef | decorated | async_stmt compound_stmt: if_stmt | while_stmt | for_stmt | try_stmt | with_stmt | funcdef | classdef | decorated | async_stmt
async_stmt: ASYNC (funcdef | with_stmt | for_stmt) async_stmt: 'async' (funcdef | with_stmt | for_stmt)
if_stmt: 'if' test ':' suite ('elif' test ':' suite)* ['else' ':' suite] if_stmt: 'if' test ':' suite ('elif' test ':' suite)* ['else' ':' suite]
while_stmt: 'while' test ':' suite ['else' ':' suite] while_stmt: 'while' test ':' suite ['else' ':' suite]
for_stmt: 'for' exprlist 'in' testlist ':' suite ['else' ':' suite] for_stmt: 'for' exprlist 'in' testlist ':' suite ['else' ':' suite]
@ -124,7 +124,7 @@ shift_expr: arith_expr (('<<'|'>>') arith_expr)*
arith_expr: term (('+'|'-') term)* arith_expr: term (('+'|'-') term)*
term: factor (('*'|'@'|'/'|'%'|'//') factor)* term: factor (('*'|'@'|'/'|'%'|'//') factor)*
factor: ('+'|'-'|'~') factor | power factor: ('+'|'-'|'~') factor | power
power: [AWAIT] atom trailer* ['**' factor] power: ['await'] atom trailer* ['**' factor]
atom: ('(' [yield_expr|testlist_gexp] ')' | atom: ('(' [yield_expr|testlist_gexp] ')' |
'[' [listmaker] ']' | '[' [listmaker] ']' |
'{' [dictsetmaker] '}' | '{' [dictsetmaker] '}' |
@ -161,7 +161,7 @@ argument: ( test [comp_for] |
star_expr ) star_expr )
comp_iter: comp_for | comp_if comp_iter: comp_for | comp_if
comp_for: [ASYNC] 'for' exprlist 'in' or_test [comp_iter] comp_for: ['async'] 'for' exprlist 'in' or_test [comp_iter]
comp_if: 'if' old_test [comp_iter] comp_if: 'if' old_test [comp_iter]
# As noted above, testlist_safe extends the syntax allowed in list # As noted above, testlist_safe extends the syntax allowed in list
@ -180,7 +180,7 @@ comp_if: 'if' old_test [comp_iter]
# #
# See https://bugs.python.org/issue27494 # See https://bugs.python.org/issue27494
old_comp_iter: old_comp_for | old_comp_if old_comp_iter: old_comp_for | old_comp_if
old_comp_for: [ASYNC] 'for' exprlist 'in' testlist_safe [old_comp_iter] old_comp_for: ['async'] 'for' exprlist 'in' testlist_safe [old_comp_iter]
old_comp_if: 'if' old_test [old_comp_iter] old_comp_if: 'if' old_test [old_comp_iter]
testlist1: test (',' test)* testlist1: test (',' test)*

View File

@ -62,10 +62,8 @@ OP = 52
COMMENT = 53 COMMENT = 53
NL = 54 NL = 54
RARROW = 55 RARROW = 55
AWAIT = 56 ERRORTOKEN = 56
ASYNC = 57 N_TOKENS = 57
ERRORTOKEN = 58
N_TOKENS = 59
NT_OFFSET = 256 NT_OFFSET = 256
#--end constants-- #--end constants--

View File

@ -234,7 +234,7 @@ class Untokenizer:
for tok in iterable: for tok in iterable:
toknum, tokval = tok[:2] toknum, tokval = tok[:2]
if toknum in (NAME, NUMBER, ASYNC, AWAIT): if toknum in (NAME, NUMBER):
tokval += ' ' tokval += ' '
if toknum == INDENT: if toknum == INDENT:
@ -380,12 +380,6 @@ def generate_tokens(readline):
contline = None contline = None
indents = [0] indents = [0]
# 'stashed' and 'async_*' are used for async/await parsing
stashed = None
async_def = False
async_def_indent = 0
async_def_nl = False
while 1: # loop over lines in stream while 1: # loop over lines in stream
try: try:
line = readline() line = readline()
@ -426,10 +420,6 @@ def generate_tokens(readline):
pos = pos + 1 pos = pos + 1
if pos == max: break if pos == max: break
if stashed:
yield stashed
stashed = None
if line[pos] in '#\r\n': # skip comments or blank lines if line[pos] in '#\r\n': # skip comments or blank lines
if line[pos] == '#': if line[pos] == '#':
comment_token = line[pos:].rstrip('\r\n') comment_token = line[pos:].rstrip('\r\n')
@ -453,18 +443,8 @@ def generate_tokens(readline):
("<tokenize>", lnum, pos, line)) ("<tokenize>", lnum, pos, line))
indents = indents[:-1] indents = indents[:-1]
if async_def and async_def_indent >= indents[-1]:
async_def = False
async_def_nl = False
async_def_indent = 0
yield (DEDENT, '', (lnum, pos), (lnum, pos), line) yield (DEDENT, '', (lnum, pos), (lnum, pos), line)
if async_def and async_def_nl and async_def_indent >= indents[-1]:
async_def = False
async_def_nl = False
async_def_indent = 0
else: # continued statement else: # continued statement
if not line: if not line:
raise TokenError("EOF in multi-line statement", (lnum, 0)) raise TokenError("EOF in multi-line statement", (lnum, 0))
@ -484,18 +464,10 @@ def generate_tokens(readline):
newline = NEWLINE newline = NEWLINE
if parenlev > 0: if parenlev > 0:
newline = NL newline = NL
elif async_def:
async_def_nl = True
if stashed:
yield stashed
stashed = None
yield (newline, token, spos, epos, line) yield (newline, token, spos, epos, line)
elif initial == '#': elif initial == '#':
assert not token.endswith("\n") assert not token.endswith("\n")
if stashed:
yield stashed
stashed = None
yield (COMMENT, token, spos, epos, line) yield (COMMENT, token, spos, epos, line)
elif token in triple_quoted: elif token in triple_quoted:
endprog = endprogs[token] endprog = endprogs[token]
@ -503,9 +475,6 @@ def generate_tokens(readline):
if endmatch: # all on one line if endmatch: # all on one line
pos = endmatch.end(0) pos = endmatch.end(0)
token = line[start:pos] token = line[start:pos]
if stashed:
yield stashed
stashed = None
yield (STRING, token, spos, (lnum, pos), line) yield (STRING, token, spos, (lnum, pos), line)
else: else:
strstart = (lnum, start) # multiple lines strstart = (lnum, start) # multiple lines
@ -523,63 +492,22 @@ def generate_tokens(readline):
contline = line contline = line
break break
else: # ordinary string else: # ordinary string
if stashed:
yield stashed
stashed = None
yield (STRING, token, spos, epos, line) yield (STRING, token, spos, epos, line)
elif initial in namechars: # ordinary name elif initial in namechars: # ordinary name
if token in ('async', 'await'): yield (NAME, token, spos, epos, line)
if async_def:
yield (ASYNC if token == 'async' else AWAIT,
token, spos, epos, line)
continue
tok = (NAME, token, spos, epos, line)
if token == 'async' and not stashed:
stashed = tok
continue
if token == 'def':
if (stashed
and stashed[0] == NAME
and stashed[1] == 'async'):
async_def = True
async_def_indent = indents[-1]
yield (ASYNC, stashed[1],
stashed[2], stashed[3],
stashed[4])
stashed = None
if stashed:
yield stashed
stashed = None
yield tok
elif initial == '\\': # continued stmt elif initial == '\\': # continued stmt
# This yield is new; needed for better idempotency: # This yield is new; needed for better idempotency:
if stashed:
yield stashed
stashed = None
yield (NL, token, spos, (lnum, pos), line) yield (NL, token, spos, (lnum, pos), line)
continued = 1 continued = 1
else: else:
if initial in '([{': parenlev = parenlev + 1 if initial in '([{': parenlev = parenlev + 1
elif initial in ')]}': parenlev = parenlev - 1 elif initial in ')]}': parenlev = parenlev - 1
if stashed:
yield stashed
stashed = None
yield (OP, token, spos, epos, line) yield (OP, token, spos, epos, line)
else: else:
yield (ERRORTOKEN, line[pos], yield (ERRORTOKEN, line[pos],
(lnum, pos), (lnum, pos+1), line) (lnum, pos), (lnum, pos+1), line)
pos = pos + 1 pos = pos + 1
if stashed:
yield stashed
stashed = None
for indent in indents[1:]: # pop remaining indent levels for indent in indents[1:]: # pop remaining indent levels
yield (DEDENT, '', (lnum, 0), (lnum, 0), '') yield (DEDENT, '', (lnum, 0), (lnum, 0), '')
yield (ENDMARKER, '', (lnum, 0), (lnum, 0), '') yield (ENDMARKER, '', (lnum, 0), (lnum, 0), '')

View File

@ -167,34 +167,34 @@ class TestAsyncAwait(GrammarTest):
async def foo(): await x async def foo(): await x
""") """)
self.invalid_syntax("await x") self.validate("await x")
self.invalid_syntax("""def foo(): self.validate("""def foo():
await x""") await x""")
self.invalid_syntax("""def foo(): self.validate("""def foo():
def foo(): pass def foo(): pass
async def foo(): pass async def foo(): pass
await x await x
""") """)
def test_async_var(self): def test_async_var(self):
self.validate("""async = 1""") self.invalid_syntax("""async = 1""")
self.validate("""await = 1""") self.invalid_syntax("""await = 1""")
self.validate("""def async(): pass""") self.invalid_syntax("""def async(): pass""")
def test_async_with(self): def test_async_with(self):
self.validate("""async def foo(): self.validate("""async def foo():
async for a in b: pass""") async for a in b: pass""")
self.invalid_syntax("""def foo(): self.validate("""def foo():
async for a in b: pass""") async for a in b: pass""")
def test_async_for(self): def test_async_for(self):
self.validate("""async def foo(): self.validate("""async def foo():
async with a: pass""") async with a: pass""")
self.invalid_syntax("""def foo(): self.validate("""def foo():
async with a: pass""") async with a: pass""")
class TestRaiseChanges(GrammarTest): class TestRaiseChanges(GrammarTest):
@ -477,3 +477,7 @@ def diff(fn, result):
os.remove("@") os.remove("@")
except OSError: except OSError:
pass pass
if __name__ == '__main__':
unittest.main()

View File

@ -1703,7 +1703,7 @@ class Helper:
# in pydoc_data/topics.py. # in pydoc_data/topics.py.
# #
# CAUTION: if you change one of these dictionaries, be sure to adapt the # CAUTION: if you change one of these dictionaries, be sure to adapt the
# list of needed labels in Doc/tools/pyspecific.py and # list of needed labels in Doc/tools/extensions/pyspecific.py and
# regenerate the pydoc_data/topics.py file by running # regenerate the pydoc_data/topics.py file by running
# make pydoc-topics # make pydoc-topics
# in Doc/ and copying the output file into the Lib/ directory. # in Doc/ and copying the output file into the Lib/ directory.
@ -1715,6 +1715,8 @@ class Helper:
'and': 'BOOLEAN', 'and': 'BOOLEAN',
'as': 'with', 'as': 'with',
'assert': ('assert', ''), 'assert': ('assert', ''),
'async': ('async', ''),
'await': ('await', ''),
'break': ('break', 'while for'), 'break': ('break', 'while for'),
'class': ('class', 'CLASSES SPECIALMETHODS'), 'class': ('class', 'CLASSES SPECIALMETHODS'),
'continue': ('continue', 'while for'), 'continue': ('continue', 'while for'),

View File

@ -91,11 +91,12 @@ classdef = 333
arglist = 334 arglist = 334
argument = 335 argument = 335
comp_iter = 336 comp_iter = 336
comp_for = 337 sync_comp_for = 337
comp_if = 338 comp_for = 338
encoding_decl = 339 comp_if = 339
yield_expr = 340 encoding_decl = 340
yield_arg = 341 yield_expr = 341
yield_arg = 342
#--end constants-- #--end constants--
sym_name = {} sym_name = {}

View File

@ -231,12 +231,6 @@ class BaseTaskTests:
with self.assertRaises(TypeError): with self.assertRaises(TypeError):
asyncio.ensure_future('ok') asyncio.ensure_future('ok')
def test_async_warning(self):
f = self.new_future(self.loop)
with self.assertWarnsRegex(DeprecationWarning,
'function is deprecated, use ensure_'):
self.assertIs(f, asyncio.async(f))
def test_get_stack(self): def test_get_stack(self):
T = None T = None

View File

@ -394,20 +394,14 @@ class AsyncBadSyntaxTest(unittest.TestCase):
] ]
for code in samples: for code in samples:
with self.subTest(code=code), self.assertWarnsRegex( with self.subTest(code=code), self.assertRaises(SyntaxError):
DeprecationWarning,
"'await' will become reserved keywords"):
compile(code, "<test>", "exec") compile(code, "<test>", "exec")
def test_badsyntax_3(self): def test_badsyntax_3(self):
with self.assertRaises(DeprecationWarning): with self.assertRaises(SyntaxError):
with warnings.catch_warnings(): compile("async = 1", "<test>", "exec")
warnings.simplefilter("error")
compile("async = 1", "<test>", "exec")
def test_goodsyntax_1(self):
# Tests for issue 24619
def test_badsyntax_4(self):
samples = [ samples = [
'''def foo(await): '''def foo(await):
async def foo(): pass async def foo(): pass
@ -454,14 +448,8 @@ class AsyncBadSyntaxTest(unittest.TestCase):
] ]
for code in samples: for code in samples:
with self.subTest(code=code): with self.subTest(code=code), self.assertRaises(SyntaxError):
loc = {} compile(code, "<test>", "exec")
with warnings.catch_warnings():
warnings.simplefilter("ignore")
exec(code, loc, loc)
self.assertEqual(loc['foo'](10), 11)
class TokenizerRegrTest(unittest.TestCase): class TokenizerRegrTest(unittest.TestCase):

View File

@ -679,16 +679,16 @@ class IllegalSyntaxTestCase(unittest.TestCase):
def test_illegal_encoding(self): def test_illegal_encoding(self):
# Illegal encoding declaration # Illegal encoding declaration
tree = \ tree = \
(339, (340,
(257, (0, ''))) (257, (0, '')))
self.check_bad_tree(tree, "missed encoding") self.check_bad_tree(tree, "missed encoding")
tree = \ tree = \
(339, (340,
(257, (0, '')), (257, (0, '')),
b'iso-8859-1') b'iso-8859-1')
self.check_bad_tree(tree, "non-string encoding") self.check_bad_tree(tree, "non-string encoding")
tree = \ tree = \
(339, (340,
(257, (0, '')), (257, (0, '')),
'\udcff') '\udcff')
with self.assertRaises(UnicodeEncodeError): with self.assertRaises(UnicodeEncodeError):

View File

@ -759,7 +759,7 @@ def"', """\
""") """)
self.check_tokenize("async def foo(): pass", """\ self.check_tokenize("async def foo(): pass", """\
ASYNC 'async' (1, 0) (1, 5) NAME 'async' (1, 0) (1, 5)
NAME 'def' (1, 6) (1, 9) NAME 'def' (1, 6) (1, 9)
NAME 'foo' (1, 10) (1, 13) NAME 'foo' (1, 10) (1, 13)
OP '(' (1, 13) (1, 14) OP '(' (1, 13) (1, 14)
@ -776,7 +776,7 @@ async def foo():
await await
async += 1 async += 1
''', """\ ''', """\
ASYNC 'async' (1, 0) (1, 5) NAME 'async' (1, 0) (1, 5)
NAME 'def' (1, 6) (1, 9) NAME 'def' (1, 6) (1, 9)
NAME 'foo' (1, 10) (1, 13) NAME 'foo' (1, 10) (1, 13)
OP '(' (1, 13) (1, 14) OP '(' (1, 13) (1, 14)
@ -787,12 +787,12 @@ async += 1
NAME 'def' (2, 2) (2, 5) NAME 'def' (2, 2) (2, 5)
NAME 'foo' (2, 6) (2, 9) NAME 'foo' (2, 6) (2, 9)
OP '(' (2, 9) (2, 10) OP '(' (2, 9) (2, 10)
AWAIT 'await' (2, 10) (2, 15) NAME 'await' (2, 10) (2, 15)
OP ')' (2, 15) (2, 16) OP ')' (2, 15) (2, 16)
OP ':' (2, 16) (2, 17) OP ':' (2, 16) (2, 17)
NEWLINE '\\n' (2, 17) (2, 18) NEWLINE '\\n' (2, 17) (2, 18)
INDENT ' ' (3, 0) (3, 4) INDENT ' ' (3, 0) (3, 4)
AWAIT 'await' (3, 4) (3, 9) NAME 'await' (3, 4) (3, 9)
OP '=' (3, 10) (3, 11) OP '=' (3, 10) (3, 11)
NUMBER '1' (3, 12) (3, 13) NUMBER '1' (3, 12) (3, 13)
NEWLINE '\\n' (3, 13) (3, 14) NEWLINE '\\n' (3, 13) (3, 14)
@ -802,7 +802,7 @@ async += 1
OP ':' (4, 6) (4, 7) OP ':' (4, 6) (4, 7)
NEWLINE '\\n' (4, 7) (4, 8) NEWLINE '\\n' (4, 7) (4, 8)
INDENT ' ' (5, 0) (5, 4) INDENT ' ' (5, 0) (5, 4)
AWAIT 'await' (5, 4) (5, 9) NAME 'await' (5, 4) (5, 9)
NEWLINE '\\n' (5, 9) (5, 10) NEWLINE '\\n' (5, 9) (5, 10)
DEDENT '' (6, 0) (6, 0) DEDENT '' (6, 0) (6, 0)
DEDENT '' (6, 0) (6, 0) DEDENT '' (6, 0) (6, 0)
@ -815,7 +815,7 @@ async += 1
self.check_tokenize('''\ self.check_tokenize('''\
async def foo(): async def foo():
async for i in 1: pass''', """\ async for i in 1: pass''', """\
ASYNC 'async' (1, 0) (1, 5) NAME 'async' (1, 0) (1, 5)
NAME 'def' (1, 6) (1, 9) NAME 'def' (1, 6) (1, 9)
NAME 'foo' (1, 10) (1, 13) NAME 'foo' (1, 10) (1, 13)
OP '(' (1, 13) (1, 14) OP '(' (1, 13) (1, 14)
@ -823,7 +823,7 @@ async def foo():
OP ':' (1, 15) (1, 16) OP ':' (1, 15) (1, 16)
NEWLINE '\\n' (1, 16) (1, 17) NEWLINE '\\n' (1, 16) (1, 17)
INDENT ' ' (2, 0) (2, 2) INDENT ' ' (2, 0) (2, 2)
ASYNC 'async' (2, 2) (2, 7) NAME 'async' (2, 2) (2, 7)
NAME 'for' (2, 8) (2, 11) NAME 'for' (2, 8) (2, 11)
NAME 'i' (2, 12) (2, 13) NAME 'i' (2, 12) (2, 13)
NAME 'in' (2, 14) (2, 16) NAME 'in' (2, 14) (2, 16)
@ -834,14 +834,14 @@ async def foo():
""") """)
self.check_tokenize('''async def foo(async): await''', """\ self.check_tokenize('''async def foo(async): await''', """\
ASYNC 'async' (1, 0) (1, 5) NAME 'async' (1, 0) (1, 5)
NAME 'def' (1, 6) (1, 9) NAME 'def' (1, 6) (1, 9)
NAME 'foo' (1, 10) (1, 13) NAME 'foo' (1, 10) (1, 13)
OP '(' (1, 13) (1, 14) OP '(' (1, 13) (1, 14)
ASYNC 'async' (1, 14) (1, 19) NAME 'async' (1, 14) (1, 19)
OP ')' (1, 19) (1, 20) OP ')' (1, 19) (1, 20)
OP ':' (1, 20) (1, 21) OP ':' (1, 20) (1, 21)
AWAIT 'await' (1, 22) (1, 27) NAME 'await' (1, 22) (1, 27)
""") """)
self.check_tokenize('''\ self.check_tokenize('''\
@ -866,7 +866,7 @@ def f():
OP ':' (3, 11) (3, 12) OP ':' (3, 11) (3, 12)
NAME 'pass' (3, 13) (3, 17) NAME 'pass' (3, 13) (3, 17)
NEWLINE '\\n' (3, 17) (3, 18) NEWLINE '\\n' (3, 17) (3, 18)
ASYNC 'async' (4, 2) (4, 7) NAME 'async' (4, 2) (4, 7)
NAME 'def' (4, 8) (4, 11) NAME 'def' (4, 8) (4, 11)
NAME 'bar' (4, 12) (4, 15) NAME 'bar' (4, 12) (4, 15)
OP '(' (4, 15) (4, 16) OP '(' (4, 15) (4, 16)
@ -888,7 +888,7 @@ async def f():
async def bar(): pass async def bar(): pass
await = 2''', """\ await = 2''', """\
ASYNC 'async' (1, 0) (1, 5) NAME 'async' (1, 0) (1, 5)
NAME 'def' (1, 6) (1, 9) NAME 'def' (1, 6) (1, 9)
NAME 'f' (1, 10) (1, 11) NAME 'f' (1, 10) (1, 11)
OP '(' (1, 11) (1, 12) OP '(' (1, 11) (1, 12)
@ -904,7 +904,7 @@ async def f():
OP ':' (3, 11) (3, 12) OP ':' (3, 11) (3, 12)
NAME 'pass' (3, 13) (3, 17) NAME 'pass' (3, 13) (3, 17)
NEWLINE '\\n' (3, 17) (3, 18) NEWLINE '\\n' (3, 17) (3, 18)
ASYNC 'async' (4, 2) (4, 7) NAME 'async' (4, 2) (4, 7)
NAME 'def' (4, 8) (4, 11) NAME 'def' (4, 8) (4, 11)
NAME 'bar' (4, 12) (4, 15) NAME 'bar' (4, 12) (4, 15)
OP '(' (4, 15) (4, 16) OP '(' (4, 15) (4, 16)
@ -913,7 +913,7 @@ async def f():
NAME 'pass' (4, 19) (4, 23) NAME 'pass' (4, 19) (4, 23)
NEWLINE '\\n' (4, 23) (4, 24) NEWLINE '\\n' (4, 23) (4, 24)
NL '\\n' (5, 0) (5, 1) NL '\\n' (5, 0) (5, 1)
AWAIT 'await' (6, 2) (6, 7) NAME 'await' (6, 2) (6, 7)
OP '=' (6, 8) (6, 9) OP '=' (6, 8) (6, 9)
NUMBER '2' (6, 10) (6, 11) NUMBER '2' (6, 10) (6, 11)
DEDENT '' (7, 0) (7, 0) DEDENT '' (7, 0) (7, 0)

View File

@ -491,12 +491,6 @@ def _tokenize(readline, encoding):
contline = None contline = None
indents = [0] indents = [0]
# 'stashed' and 'async_*' are used for async/await parsing
stashed = None
async_def = False
async_def_indent = 0
async_def_nl = False
if encoding is not None: if encoding is not None:
if encoding == "utf-8-sig": if encoding == "utf-8-sig":
# BOM will already have been stripped. # BOM will already have been stripped.
@ -571,18 +565,8 @@ def _tokenize(readline, encoding):
("<tokenize>", lnum, pos, line)) ("<tokenize>", lnum, pos, line))
indents = indents[:-1] indents = indents[:-1]
if async_def and async_def_indent >= indents[-1]:
async_def = False
async_def_nl = False
async_def_indent = 0
yield TokenInfo(DEDENT, '', (lnum, pos), (lnum, pos), line) yield TokenInfo(DEDENT, '', (lnum, pos), (lnum, pos), line)
if async_def and async_def_nl and async_def_indent >= indents[-1]:
async_def = False
async_def_nl = False
async_def_indent = 0
else: # continued statement else: # continued statement
if not line: if not line:
raise TokenError("EOF in multi-line statement", (lnum, 0)) raise TokenError("EOF in multi-line statement", (lnum, 0))
@ -601,21 +585,13 @@ def _tokenize(readline, encoding):
(initial == '.' and token != '.' and token != '...')): (initial == '.' and token != '.' and token != '...')):
yield TokenInfo(NUMBER, token, spos, epos, line) yield TokenInfo(NUMBER, token, spos, epos, line)
elif initial in '\r\n': elif initial in '\r\n':
if stashed:
yield stashed
stashed = None
if parenlev > 0: if parenlev > 0:
yield TokenInfo(NL, token, spos, epos, line) yield TokenInfo(NL, token, spos, epos, line)
else: else:
yield TokenInfo(NEWLINE, token, spos, epos, line) yield TokenInfo(NEWLINE, token, spos, epos, line)
if async_def:
async_def_nl = True
elif initial == '#': elif initial == '#':
assert not token.endswith("\n") assert not token.endswith("\n")
if stashed:
yield stashed
stashed = None
yield TokenInfo(COMMENT, token, spos, epos, line) yield TokenInfo(COMMENT, token, spos, epos, line)
elif token in triple_quoted: elif token in triple_quoted:
@ -662,36 +638,7 @@ def _tokenize(readline, encoding):
yield TokenInfo(STRING, token, spos, epos, line) yield TokenInfo(STRING, token, spos, epos, line)
elif initial.isidentifier(): # ordinary name elif initial.isidentifier(): # ordinary name
if token in ('async', 'await'): yield TokenInfo(NAME, token, spos, epos, line)
if async_def:
yield TokenInfo(
ASYNC if token == 'async' else AWAIT,
token, spos, epos, line)
continue
tok = TokenInfo(NAME, token, spos, epos, line)
if token == 'async' and not stashed:
stashed = tok
continue
if token == 'def':
if (stashed
and stashed.type == NAME
and stashed.string == 'async'):
async_def = True
async_def_indent = indents[-1]
yield TokenInfo(ASYNC, stashed.string,
stashed.start, stashed.end,
stashed.line)
stashed = None
if stashed:
yield stashed
stashed = None
yield tok
elif initial == '\\': # continued stmt elif initial == '\\': # continued stmt
continued = 1 continued = 1
else: else:
@ -699,19 +646,12 @@ def _tokenize(readline, encoding):
parenlev += 1 parenlev += 1
elif initial in ')]}': elif initial in ')]}':
parenlev -= 1 parenlev -= 1
if stashed:
yield stashed
stashed = None
yield TokenInfo(OP, token, spos, epos, line) yield TokenInfo(OP, token, spos, epos, line)
else: else:
yield TokenInfo(ERRORTOKEN, line[pos], yield TokenInfo(ERRORTOKEN, line[pos],
(lnum, pos), (lnum, pos+1), line) (lnum, pos), (lnum, pos+1), line)
pos += 1 pos += 1
if stashed:
yield stashed
stashed = None
for indent in indents[1:]: # pop remaining indent levels for indent in indents[1:]: # pop remaining indent levels
yield TokenInfo(DEDENT, '', (lnum, 0), (lnum, 0), '') yield TokenInfo(DEDENT, '', (lnum, 0), (lnum, 0), '')
yield TokenInfo(ENDMARKER, '', (lnum, 0), (lnum, 0), '') yield TokenInfo(ENDMARKER, '', (lnum, 0), (lnum, 0), '')

View File

@ -0,0 +1 @@
Make ``async`` and ``await`` proper keywords, as specified in PEP 492.

View File

@ -103,8 +103,6 @@ const char *_PyParser_TokenNames[] = {
"ELLIPSIS", "ELLIPSIS",
/* This table must match the #defines in token.h! */ /* This table must match the #defines in token.h! */
"OP", "OP",
"AWAIT",
"ASYNC",
"<ERRORTOKEN>", "<ERRORTOKEN>",
"COMMENT", "COMMENT",
"NL", "NL",
@ -151,10 +149,6 @@ tok_new(void)
tok->decoding_buffer = NULL; tok->decoding_buffer = NULL;
#endif #endif
tok->async_def = 0;
tok->async_def_indent = 0;
tok->async_def_nl = 0;
return tok; return tok;
} }
@ -1471,21 +1465,6 @@ tok_get(struct tok_state *tok, char **p_start, char **p_end)
} }
} }
if (tok->async_def
&& !blankline
&& tok->level == 0
/* There was a NEWLINE after ASYNC DEF,
so we're past the signature. */
&& tok->async_def_nl
/* Current indentation level is less than where
the async function was defined */
&& tok->async_def_indent >= tok->indent)
{
tok->async_def = 0;
tok->async_def_indent = 0;
tok->async_def_nl = 0;
}
again: again:
tok->start = NULL; tok->start = NULL;
/* Skip spaces */ /* Skip spaces */
@ -1550,43 +1529,6 @@ tok_get(struct tok_state *tok, char **p_start, char **p_end)
*p_start = tok->start; *p_start = tok->start;
*p_end = tok->cur; *p_end = tok->cur;
/* async/await parsing block. */
if (tok->cur - tok->start == 5) {
/* Current token length is 5. */
if (tok->async_def) {
/* We're inside an 'async def' function. */
if (memcmp(tok->start, "async", 5) == 0) {
return ASYNC;
}
if (memcmp(tok->start, "await", 5) == 0) {
return AWAIT;
}
}
else if (memcmp(tok->start, "async", 5) == 0) {
/* The current token is 'async'.
Look ahead one token.*/
struct tok_state ahead_tok;
char *ahead_tok_start = NULL, *ahead_tok_end = NULL;
int ahead_tok_kind;
memcpy(&ahead_tok, tok, sizeof(ahead_tok));
ahead_tok_kind = tok_get(&ahead_tok, &ahead_tok_start,
&ahead_tok_end);
if (ahead_tok_kind == NAME
&& ahead_tok.cur - ahead_tok.start == 3
&& memcmp(ahead_tok.start, "def", 3) == 0)
{
/* The next token is going to be 'def', so instead of
returning 'async' NAME token, we return ASYNC. */
tok->async_def_indent = tok->indent;
tok->async_def = 1;
return ASYNC;
}
}
}
return NAME; return NAME;
} }
@ -1599,11 +1541,6 @@ tok_get(struct tok_state *tok, char **p_start, char **p_end)
*p_start = tok->start; *p_start = tok->start;
*p_end = tok->cur - 1; /* Leave '\n' out of the string */ *p_end = tok->cur - 1; /* Leave '\n' out of the string */
tok->cont_line = 0; tok->cont_line = 0;
if (tok->async_def) {
/* We're somewhere inside an 'async def' function, and
we've encountered a NEWLINE after its signature. */
tok->async_def_nl = 1;
}
return NEWLINE; return NEWLINE;
} }

View File

@ -65,13 +65,6 @@ struct tok_state {
const char* enc; /* Encoding for the current str. */ const char* enc; /* Encoding for the current str. */
const char* str; const char* str;
const char* input; /* Tokenizer's newline translated copy of the string. */ const char* input; /* Tokenizer's newline translated copy of the string. */
/* async/await related fields; can be removed in 3.7 when async and await
become normal keywords. */
int async_def; /* =1 if tokens are inside an 'async def' body. */
int async_def_indent; /* Indentation level of the outermost 'async def'. */
int async_def_nl; /* =1 if the outermost 'async def' had at least one
NEWLINE token after it. */
}; };
extern struct tok_state *PyTokenizer_FromString(const char *, int); extern struct tok_state *PyTokenizer_FromString(const char *, int);

View File

@ -949,28 +949,6 @@ forbidden_name(struct compiling *c, identifier name, const node *n,
ast_error(c, n, "assignment to keyword"); ast_error(c, n, "assignment to keyword");
return 1; return 1;
} }
if (_PyUnicode_EqualToASCIIString(name, "async") ||
_PyUnicode_EqualToASCIIString(name, "await"))
{
PyObject *message = PyUnicode_FromString(
"'async' and 'await' will become reserved keywords"
" in Python 3.7");
int ret;
if (message == NULL) {
return 1;
}
ret = PyErr_WarnExplicitObject(
PyExc_DeprecationWarning,
message,
c->c_filename,
LINENO(n),
NULL,
NULL);
Py_DECREF(message);
if (ret < 0) {
return 1;
}
}
if (full_checks) { if (full_checks) {
const char * const *p; const char * const *p;
for (p = FORBIDDEN; *p; p++) { for (p = FORBIDDEN; *p; p++) {
@ -1642,9 +1620,10 @@ ast_for_funcdef_impl(struct compiling *c, const node *n,
static stmt_ty static stmt_ty
ast_for_async_funcdef(struct compiling *c, const node *n, asdl_seq *decorator_seq) ast_for_async_funcdef(struct compiling *c, const node *n, asdl_seq *decorator_seq)
{ {
/* async_funcdef: ASYNC funcdef */ /* async_funcdef: 'async' funcdef */
REQ(n, async_funcdef); REQ(n, async_funcdef);
REQ(CHILD(n, 0), ASYNC); REQ(CHILD(n, 0), NAME);
assert(strcmp(STR(CHILD(n, 0)), "async") == 0);
REQ(CHILD(n, 1), funcdef); REQ(CHILD(n, 1), funcdef);
return ast_for_funcdef_impl(c, CHILD(n, 1), decorator_seq, return ast_for_funcdef_impl(c, CHILD(n, 1), decorator_seq,
@ -1663,9 +1642,10 @@ ast_for_funcdef(struct compiling *c, const node *n, asdl_seq *decorator_seq)
static stmt_ty static stmt_ty
ast_for_async_stmt(struct compiling *c, const node *n) ast_for_async_stmt(struct compiling *c, const node *n)
{ {
/* async_stmt: ASYNC (funcdef | with_stmt | for_stmt) */ /* async_stmt: 'async' (funcdef | with_stmt | for_stmt) */
REQ(n, async_stmt); REQ(n, async_stmt);
REQ(CHILD(n, 0), ASYNC); REQ(CHILD(n, 0), NAME);
assert(strcmp(STR(CHILD(n, 0)), "async") == 0);
switch (TYPE(CHILD(n, 1))) { switch (TYPE(CHILD(n, 1))) {
case funcdef: case funcdef:
@ -1778,17 +1758,23 @@ static int
count_comp_fors(struct compiling *c, const node *n) count_comp_fors(struct compiling *c, const node *n)
{ {
int n_fors = 0; int n_fors = 0;
int is_async;
count_comp_for: count_comp_for:
is_async = 0;
n_fors++; n_fors++;
REQ(n, comp_for); REQ(n, comp_for);
if (TYPE(CHILD(n, 0)) == ASYNC) { if (NCH(n) == 2) {
is_async = 1; REQ(CHILD(n, 0), NAME);
assert(strcmp(STR(CHILD(n, 0)), "async") == 0);
n = CHILD(n, 1);
} }
if (NCH(n) == (5 + is_async)) { else if (NCH(n) == 1) {
n = CHILD(n, 4 + is_async); n = CHILD(n, 0);
}
else {
goto error;
}
if (NCH(n) == (5)) {
n = CHILD(n, 4);
} }
else { else {
return n_fors; return n_fors;
@ -1807,6 +1793,7 @@ count_comp_fors(struct compiling *c, const node *n)
return n_fors; return n_fors;
} }
error:
/* Should never be reached */ /* Should never be reached */
PyErr_SetString(PyExc_SystemError, PyErr_SetString(PyExc_SystemError,
"logic error in count_comp_fors"); "logic error in count_comp_fors");
@ -1855,19 +1842,27 @@ ast_for_comprehension(struct compiling *c, const node *n)
asdl_seq *t; asdl_seq *t;
expr_ty expression, first; expr_ty expression, first;
node *for_ch; node *for_ch;
node *sync_n;
int is_async = 0; int is_async = 0;
REQ(n, comp_for); REQ(n, comp_for);
if (TYPE(CHILD(n, 0)) == ASYNC) { if (NCH(n) == 2) {
is_async = 1; is_async = 1;
REQ(CHILD(n, 0), NAME);
assert(strcmp(STR(CHILD(n, 0)), "async") == 0);
sync_n = CHILD(n, 1);
} }
else {
sync_n = CHILD(n, 0);
}
REQ(sync_n, sync_comp_for);
for_ch = CHILD(n, 1 + is_async); for_ch = CHILD(sync_n, 1);
t = ast_for_exprlist(c, for_ch, Store); t = ast_for_exprlist(c, for_ch, Store);
if (!t) if (!t)
return NULL; return NULL;
expression = ast_for_expr(c, CHILD(n, 3 + is_async)); expression = ast_for_expr(c, CHILD(sync_n, 3));
if (!expression) if (!expression)
return NULL; return NULL;
@ -1884,11 +1879,11 @@ ast_for_comprehension(struct compiling *c, const node *n)
if (!comp) if (!comp)
return NULL; return NULL;
if (NCH(n) == (5 + is_async)) { if (NCH(sync_n) == 5) {
int j, n_ifs; int j, n_ifs;
asdl_seq *ifs; asdl_seq *ifs;
n = CHILD(n, 4 + is_async); n = CHILD(sync_n, 4);
n_ifs = count_comp_ifs(c, n); n_ifs = count_comp_ifs(c, n);
if (n_ifs == -1) if (n_ifs == -1)
return NULL; return NULL;
@ -2470,7 +2465,7 @@ ast_for_atom_expr(struct compiling *c, const node *n)
REQ(n, atom_expr); REQ(n, atom_expr);
nch = NCH(n); nch = NCH(n);
if (TYPE(CHILD(n, 0)) == AWAIT) { if (TYPE(CHILD(n, 0)) == NAME && strcmp(STR(CHILD(n, 0)), "await") == 0) {
start = 1; start = 1;
assert(nch > 1); assert(nch > 1);
} }
@ -2497,7 +2492,7 @@ ast_for_atom_expr(struct compiling *c, const node *n)
} }
if (start) { if (start) {
/* there was an AWAIT */ /* there was an 'await' */
return Await(e, LINENO(n), n->n_col_offset, c->c_arena); return Await(e, LINENO(n), n->n_col_offset, c->c_arena);
} }
else { else {
@ -2562,7 +2557,7 @@ ast_for_expr(struct compiling *c, const node *n)
term: factor (('*'|'@'|'/'|'%'|'//') factor)* term: factor (('*'|'@'|'/'|'%'|'//') factor)*
factor: ('+'|'-'|'~') factor | power factor: ('+'|'-'|'~') factor | power
power: atom_expr ['**' factor] power: atom_expr ['**' factor]
atom_expr: [AWAIT] atom trailer* atom_expr: ['await'] atom trailer*
yield_expr: 'yield' [yield_arg] yield_expr: 'yield' [yield_arg]
*/ */

View File

@ -1812,272 +1812,284 @@ static state states_80[2] = {
{2, arcs_80_0}, {2, arcs_80_0},
{1, arcs_80_1}, {1, arcs_80_1},
}; };
static arc arcs_81_0[2] = { static arc arcs_81_0[1] = {
{21, 1}, {101, 1},
{101, 2},
}; };
static arc arcs_81_1[1] = { static arc arcs_81_1[1] = {
{101, 2}, {66, 2},
}; };
static arc arcs_81_2[1] = { static arc arcs_81_2[1] = {
{66, 3}, {102, 3},
}; };
static arc arcs_81_3[1] = { static arc arcs_81_3[1] = {
{102, 4}, {112, 4},
}; };
static arc arcs_81_4[1] = { static arc arcs_81_4[2] = {
{112, 5}, {171, 5},
{0, 4},
}; };
static arc arcs_81_5[2] = { static arc arcs_81_5[1] = {
{171, 6},
{0, 5}, {0, 5},
}; };
static arc arcs_81_6[1] = { static state states_81[6] = {
{0, 6}, {1, arcs_81_0},
};
static state states_81[7] = {
{2, arcs_81_0},
{1, arcs_81_1}, {1, arcs_81_1},
{1, arcs_81_2}, {1, arcs_81_2},
{1, arcs_81_3}, {1, arcs_81_3},
{1, arcs_81_4}, {2, arcs_81_4},
{2, arcs_81_5}, {1, arcs_81_5},
{1, arcs_81_6},
}; };
static arc arcs_82_0[1] = { static arc arcs_82_0[2] = {
{97, 1}, {21, 1},
{173, 2},
}; };
static arc arcs_82_1[1] = { static arc arcs_82_1[1] = {
{173, 2},
};
static arc arcs_82_2[1] = {
{0, 2},
};
static state states_82[3] = {
{2, arcs_82_0},
{1, arcs_82_1},
{1, arcs_82_2},
};
static arc arcs_83_0[1] = {
{97, 1},
};
static arc arcs_83_1[1] = {
{114, 2}, {114, 2},
}; };
static arc arcs_82_2[2] = { static arc arcs_83_2[2] = {
{171, 3}, {171, 3},
{0, 2}, {0, 2},
}; };
static arc arcs_82_3[1] = { static arc arcs_83_3[1] = {
{0, 3}, {0, 3},
}; };
static state states_82[4] = { static state states_83[4] = {
{1, arcs_82_0},
{1, arcs_82_1},
{2, arcs_82_2},
{1, arcs_82_3},
};
static arc arcs_83_0[1] = {
{23, 1},
};
static arc arcs_83_1[1] = {
{0, 1},
};
static state states_83[2] = {
{1, arcs_83_0}, {1, arcs_83_0},
{1, arcs_83_1}, {1, arcs_83_1},
{2, arcs_83_2},
{1, arcs_83_3},
}; };
static arc arcs_84_0[1] = { static arc arcs_84_0[1] = {
{174, 1}, {23, 1},
}; };
static arc arcs_84_1[2] = { static arc arcs_84_1[1] = {
{175, 2},
{0, 1}, {0, 1},
}; };
static arc arcs_84_2[1] = { static state states_84[2] = {
{0, 2},
};
static state states_84[3] = {
{1, arcs_84_0}, {1, arcs_84_0},
{2, arcs_84_1}, {1, arcs_84_1},
{1, arcs_84_2},
}; };
static arc arcs_85_0[2] = { static arc arcs_85_0[1] = {
{77, 1}, {175, 1},
{9, 2},
}; };
static arc arcs_85_1[1] = { static arc arcs_85_1[2] = {
{26, 2}, {176, 2},
{0, 1},
}; };
static arc arcs_85_2[1] = { static arc arcs_85_2[1] = {
{0, 2}, {0, 2},
}; };
static state states_85[3] = { static state states_85[3] = {
{2, arcs_85_0}, {1, arcs_85_0},
{1, arcs_85_1}, {2, arcs_85_1},
{1, arcs_85_2}, {1, arcs_85_2},
}; };
static dfa dfas[86] = { static arc arcs_86_0[2] = {
{256, "single_input", 0, 3, states_0, {77, 1},
"\004\050\340\000\002\000\000\000\012\076\011\007\262\004\020\002\000\300\220\050\037\102"}, {9, 2},
{257, "file_input", 0, 2, states_1,
"\204\050\340\000\002\000\000\000\012\076\011\007\262\004\020\002\000\300\220\050\037\102"},
{258, "eval_input", 0, 3, states_2,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000"},
{259, "decorator", 0, 7, states_3,
"\000\010\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{260, "decorators", 0, 2, states_4,
"\000\010\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{261, "decorated", 0, 3, states_5,
"\000\010\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{262, "async_funcdef", 0, 3, states_6,
"\000\000\040\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{263, "funcdef", 0, 8, states_7,
"\000\000\100\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{264, "parameters", 0, 4, states_8,
"\000\040\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{265, "typedargslist", 0, 19, states_9,
"\000\000\200\000\006\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{266, "tfpdef", 0, 4, states_10,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{267, "varargslist", 0, 19, states_11,
"\000\000\200\000\006\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{268, "vfpdef", 0, 2, states_12,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{269, "stmt", 0, 2, states_13,
"\000\050\340\000\002\000\000\000\012\076\011\007\262\004\020\002\000\300\220\050\037\102"},
{270, "simple_stmt", 0, 4, states_14,
"\000\040\200\000\002\000\000\000\012\076\011\007\000\000\020\002\000\300\220\050\037\100"},
{271, "small_stmt", 0, 2, states_15,
"\000\040\200\000\002\000\000\000\012\076\011\007\000\000\020\002\000\300\220\050\037\100"},
{272, "expr_stmt", 0, 6, states_16,
"\000\040\200\000\002\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000"},
{273, "annassign", 0, 5, states_17,
"\000\000\000\010\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{274, "testlist_star_expr", 0, 3, states_18,
"\000\040\200\000\002\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000"},
{275, "augassign", 0, 2, states_19,
"\000\000\000\000\000\000\360\377\001\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{276, "del_stmt", 0, 3, states_20,
"\000\000\000\000\000\000\000\000\002\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{277, "pass_stmt", 0, 2, states_21,
"\000\000\000\000\000\000\000\000\010\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{278, "flow_stmt", 0, 2, states_22,
"\000\000\000\000\000\000\000\000\000\036\000\000\000\000\000\000\000\000\000\000\000\100"},
{279, "break_stmt", 0, 2, states_23,
"\000\000\000\000\000\000\000\000\000\002\000\000\000\000\000\000\000\000\000\000\000\000"},
{280, "continue_stmt", 0, 2, states_24,
"\000\000\000\000\000\000\000\000\000\004\000\000\000\000\000\000\000\000\000\000\000\000"},
{281, "return_stmt", 0, 3, states_25,
"\000\000\000\000\000\000\000\000\000\010\000\000\000\000\000\000\000\000\000\000\000\000"},
{282, "yield_stmt", 0, 2, states_26,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\100"},
{283, "raise_stmt", 0, 5, states_27,
"\000\000\000\000\000\000\000\000\000\020\000\000\000\000\000\000\000\000\000\000\000\000"},
{284, "import_stmt", 0, 2, states_28,
"\000\000\000\000\000\000\000\000\000\040\001\000\000\000\000\000\000\000\000\000\000\000"},
{285, "import_name", 0, 3, states_29,
"\000\000\000\000\000\000\000\000\000\000\001\000\000\000\000\000\000\000\000\000\000\000"},
{286, "import_from", 0, 8, states_30,
"\000\000\000\000\000\000\000\000\000\040\000\000\000\000\000\000\000\000\000\000\000\000"},
{287, "import_as_name", 0, 4, states_31,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{288, "dotted_as_name", 0, 4, states_32,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{289, "import_as_names", 0, 3, states_33,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{290, "dotted_as_names", 0, 2, states_34,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{291, "dotted_name", 0, 2, states_35,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{292, "global_stmt", 0, 3, states_36,
"\000\000\000\000\000\000\000\000\000\000\000\001\000\000\000\000\000\000\000\000\000\000"},
{293, "nonlocal_stmt", 0, 3, states_37,
"\000\000\000\000\000\000\000\000\000\000\000\002\000\000\000\000\000\000\000\000\000\000"},
{294, "assert_stmt", 0, 5, states_38,
"\000\000\000\000\000\000\000\000\000\000\000\004\000\000\000\000\000\000\000\000\000\000"},
{295, "compound_stmt", 0, 2, states_39,
"\000\010\140\000\000\000\000\000\000\000\000\000\262\004\000\000\000\000\000\000\000\002"},
{296, "async_stmt", 0, 3, states_40,
"\000\000\040\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{297, "if_stmt", 0, 8, states_41,
"\000\000\000\000\000\000\000\000\000\000\000\000\002\000\000\000\000\000\000\000\000\000"},
{298, "while_stmt", 0, 8, states_42,
"\000\000\000\000\000\000\000\000\000\000\000\000\020\000\000\000\000\000\000\000\000\000"},
{299, "for_stmt", 0, 10, states_43,
"\000\000\000\000\000\000\000\000\000\000\000\000\040\000\000\000\000\000\000\000\000\000"},
{300, "try_stmt", 0, 13, states_44,
"\000\000\000\000\000\000\000\000\000\000\000\000\200\000\000\000\000\000\000\000\000\000"},
{301, "with_stmt", 0, 5, states_45,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\004\000\000\000\000\000\000\000\000"},
{302, "with_item", 0, 4, states_46,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000"},
{303, "except_clause", 0, 5, states_47,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\040\000\000\000\000\000\000\000\000"},
{304, "suite", 0, 5, states_48,
"\004\040\200\000\002\000\000\000\012\076\011\007\000\000\020\002\000\300\220\050\037\100"},
{305, "test", 0, 6, states_49,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000"},
{306, "test_nocond", 0, 2, states_50,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000"},
{307, "lambdef", 0, 5, states_51,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\000\020\000\000\000\000\000\000\000"},
{308, "lambdef_nocond", 0, 5, states_52,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\000\020\000\000\000\000\000\000\000"},
{309, "or_test", 0, 2, states_53,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\002\000\300\220\050\037\000"},
{310, "and_test", 0, 2, states_54,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\002\000\300\220\050\037\000"},
{311, "not_test", 0, 3, states_55,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\002\000\300\220\050\037\000"},
{312, "comparison", 0, 2, states_56,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000"},
{313, "comp_op", 0, 4, states_57,
"\000\000\000\000\000\000\000\000\000\000\000\000\100\000\000\362\017\000\000\000\000\000"},
{314, "star_expr", 0, 3, states_58,
"\000\000\000\000\002\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{315, "expr", 0, 2, states_59,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000"},
{316, "xor_expr", 0, 2, states_60,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000"},
{317, "and_expr", 0, 2, states_61,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000"},
{318, "shift_expr", 0, 2, states_62,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000"},
{319, "arith_expr", 0, 2, states_63,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000"},
{320, "term", 0, 2, states_64,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000"},
{321, "factor", 0, 3, states_65,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000"},
{322, "power", 0, 4, states_66,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\000\200\050\037\000"},
{323, "atom_expr", 0, 3, states_67,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\000\200\050\037\000"},
{324, "atom", 0, 9, states_68,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\000\000\050\037\000"},
{325, "testlist_comp", 0, 5, states_69,
"\000\040\200\000\002\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000"},
{326, "trailer", 0, 7, states_70,
"\000\040\000\000\000\000\000\000\000\000\004\000\000\000\000\000\000\000\000\010\000\000"},
{327, "subscriptlist", 0, 3, states_71,
"\000\040\200\010\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000"},
{328, "subscript", 0, 5, states_72,
"\000\040\200\010\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000"},
{329, "sliceop", 0, 3, states_73,
"\000\000\000\010\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{330, "exprlist", 0, 3, states_74,
"\000\040\200\000\002\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000"},
{331, "testlist", 0, 3, states_75,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000"},
{332, "dictorsetmaker", 0, 14, states_76,
"\000\040\200\000\006\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000"},
{333, "classdef", 0, 8, states_77,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\002"},
{334, "arglist", 0, 3, states_78,
"\000\040\200\000\006\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000"},
{335, "argument", 0, 4, states_79,
"\000\040\200\000\006\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000"},
{336, "comp_iter", 0, 2, states_80,
"\000\000\040\000\000\000\000\000\000\000\000\000\042\000\000\000\000\000\000\000\000\000"},
{337, "comp_for", 0, 7, states_81,
"\000\000\040\000\000\000\000\000\000\000\000\000\040\000\000\000\000\000\000\000\000\000"},
{338, "comp_if", 0, 4, states_82,
"\000\000\000\000\000\000\000\000\000\000\000\000\002\000\000\000\000\000\000\000\000\000"},
{339, "encoding_decl", 0, 2, states_83,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{340, "yield_expr", 0, 3, states_84,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\100"},
{341, "yield_arg", 0, 3, states_85,
"\000\040\200\000\000\000\000\000\000\040\010\000\000\000\020\002\000\300\220\050\037\000"},
}; };
static label labels[176] = { static arc arcs_86_1[1] = {
{26, 2},
};
static arc arcs_86_2[1] = {
{0, 2},
};
static state states_86[3] = {
{2, arcs_86_0},
{1, arcs_86_1},
{1, arcs_86_2},
};
static dfa dfas[87] = {
{256, "single_input", 0, 3, states_0,
"\004\050\340\000\002\000\000\000\012\076\011\007\262\004\020\002\000\300\220\050\037\202\000"},
{257, "file_input", 0, 2, states_1,
"\204\050\340\000\002\000\000\000\012\076\011\007\262\004\020\002\000\300\220\050\037\202\000"},
{258, "eval_input", 0, 3, states_2,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000\000"},
{259, "decorator", 0, 7, states_3,
"\000\010\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{260, "decorators", 0, 2, states_4,
"\000\010\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{261, "decorated", 0, 3, states_5,
"\000\010\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{262, "async_funcdef", 0, 3, states_6,
"\000\000\040\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{263, "funcdef", 0, 8, states_7,
"\000\000\100\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{264, "parameters", 0, 4, states_8,
"\000\040\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{265, "typedargslist", 0, 19, states_9,
"\000\000\200\000\006\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{266, "tfpdef", 0, 4, states_10,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{267, "varargslist", 0, 19, states_11,
"\000\000\200\000\006\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{268, "vfpdef", 0, 2, states_12,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{269, "stmt", 0, 2, states_13,
"\000\050\340\000\002\000\000\000\012\076\011\007\262\004\020\002\000\300\220\050\037\202\000"},
{270, "simple_stmt", 0, 4, states_14,
"\000\040\200\000\002\000\000\000\012\076\011\007\000\000\020\002\000\300\220\050\037\200\000"},
{271, "small_stmt", 0, 2, states_15,
"\000\040\200\000\002\000\000\000\012\076\011\007\000\000\020\002\000\300\220\050\037\200\000"},
{272, "expr_stmt", 0, 6, states_16,
"\000\040\200\000\002\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000\000"},
{273, "annassign", 0, 5, states_17,
"\000\000\000\010\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{274, "testlist_star_expr", 0, 3, states_18,
"\000\040\200\000\002\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000\000"},
{275, "augassign", 0, 2, states_19,
"\000\000\000\000\000\000\360\377\001\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{276, "del_stmt", 0, 3, states_20,
"\000\000\000\000\000\000\000\000\002\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{277, "pass_stmt", 0, 2, states_21,
"\000\000\000\000\000\000\000\000\010\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{278, "flow_stmt", 0, 2, states_22,
"\000\000\000\000\000\000\000\000\000\036\000\000\000\000\000\000\000\000\000\000\000\200\000"},
{279, "break_stmt", 0, 2, states_23,
"\000\000\000\000\000\000\000\000\000\002\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{280, "continue_stmt", 0, 2, states_24,
"\000\000\000\000\000\000\000\000\000\004\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{281, "return_stmt", 0, 3, states_25,
"\000\000\000\000\000\000\000\000\000\010\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{282, "yield_stmt", 0, 2, states_26,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\200\000"},
{283, "raise_stmt", 0, 5, states_27,
"\000\000\000\000\000\000\000\000\000\020\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{284, "import_stmt", 0, 2, states_28,
"\000\000\000\000\000\000\000\000\000\040\001\000\000\000\000\000\000\000\000\000\000\000\000"},
{285, "import_name", 0, 3, states_29,
"\000\000\000\000\000\000\000\000\000\000\001\000\000\000\000\000\000\000\000\000\000\000\000"},
{286, "import_from", 0, 8, states_30,
"\000\000\000\000\000\000\000\000\000\040\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{287, "import_as_name", 0, 4, states_31,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{288, "dotted_as_name", 0, 4, states_32,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{289, "import_as_names", 0, 3, states_33,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{290, "dotted_as_names", 0, 2, states_34,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{291, "dotted_name", 0, 2, states_35,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{292, "global_stmt", 0, 3, states_36,
"\000\000\000\000\000\000\000\000\000\000\000\001\000\000\000\000\000\000\000\000\000\000\000"},
{293, "nonlocal_stmt", 0, 3, states_37,
"\000\000\000\000\000\000\000\000\000\000\000\002\000\000\000\000\000\000\000\000\000\000\000"},
{294, "assert_stmt", 0, 5, states_38,
"\000\000\000\000\000\000\000\000\000\000\000\004\000\000\000\000\000\000\000\000\000\000\000"},
{295, "compound_stmt", 0, 2, states_39,
"\000\010\140\000\000\000\000\000\000\000\000\000\262\004\000\000\000\000\000\000\000\002\000"},
{296, "async_stmt", 0, 3, states_40,
"\000\000\040\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{297, "if_stmt", 0, 8, states_41,
"\000\000\000\000\000\000\000\000\000\000\000\000\002\000\000\000\000\000\000\000\000\000\000"},
{298, "while_stmt", 0, 8, states_42,
"\000\000\000\000\000\000\000\000\000\000\000\000\020\000\000\000\000\000\000\000\000\000\000"},
{299, "for_stmt", 0, 10, states_43,
"\000\000\000\000\000\000\000\000\000\000\000\000\040\000\000\000\000\000\000\000\000\000\000"},
{300, "try_stmt", 0, 13, states_44,
"\000\000\000\000\000\000\000\000\000\000\000\000\200\000\000\000\000\000\000\000\000\000\000"},
{301, "with_stmt", 0, 5, states_45,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\004\000\000\000\000\000\000\000\000\000"},
{302, "with_item", 0, 4, states_46,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000\000"},
{303, "except_clause", 0, 5, states_47,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\040\000\000\000\000\000\000\000\000\000"},
{304, "suite", 0, 5, states_48,
"\004\040\200\000\002\000\000\000\012\076\011\007\000\000\020\002\000\300\220\050\037\200\000"},
{305, "test", 0, 6, states_49,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000\000"},
{306, "test_nocond", 0, 2, states_50,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000\000"},
{307, "lambdef", 0, 5, states_51,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\000\020\000\000\000\000\000\000\000\000"},
{308, "lambdef_nocond", 0, 5, states_52,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\000\020\000\000\000\000\000\000\000\000"},
{309, "or_test", 0, 2, states_53,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\002\000\300\220\050\037\000\000"},
{310, "and_test", 0, 2, states_54,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\002\000\300\220\050\037\000\000"},
{311, "not_test", 0, 3, states_55,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\002\000\300\220\050\037\000\000"},
{312, "comparison", 0, 2, states_56,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000\000"},
{313, "comp_op", 0, 4, states_57,
"\000\000\000\000\000\000\000\000\000\000\000\000\100\000\000\362\017\000\000\000\000\000\000"},
{314, "star_expr", 0, 3, states_58,
"\000\000\000\000\002\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{315, "expr", 0, 2, states_59,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000\000"},
{316, "xor_expr", 0, 2, states_60,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000\000"},
{317, "and_expr", 0, 2, states_61,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000\000"},
{318, "shift_expr", 0, 2, states_62,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000\000"},
{319, "arith_expr", 0, 2, states_63,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000\000"},
{320, "term", 0, 2, states_64,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000\000"},
{321, "factor", 0, 3, states_65,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000\000"},
{322, "power", 0, 4, states_66,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\000\200\050\037\000\000"},
{323, "atom_expr", 0, 3, states_67,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\000\200\050\037\000\000"},
{324, "atom", 0, 9, states_68,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\000\000\000\000\000\050\037\000\000"},
{325, "testlist_comp", 0, 5, states_69,
"\000\040\200\000\002\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000\000"},
{326, "trailer", 0, 7, states_70,
"\000\040\000\000\000\000\000\000\000\000\004\000\000\000\000\000\000\000\000\010\000\000\000"},
{327, "subscriptlist", 0, 3, states_71,
"\000\040\200\010\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000\000"},
{328, "subscript", 0, 5, states_72,
"\000\040\200\010\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000\000"},
{329, "sliceop", 0, 3, states_73,
"\000\000\000\010\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{330, "exprlist", 0, 3, states_74,
"\000\040\200\000\002\000\000\000\000\000\010\000\000\000\000\000\000\300\220\050\037\000\000"},
{331, "testlist", 0, 3, states_75,
"\000\040\200\000\000\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000\000"},
{332, "dictorsetmaker", 0, 14, states_76,
"\000\040\200\000\006\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000\000"},
{333, "classdef", 0, 8, states_77,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\002\000"},
{334, "arglist", 0, 3, states_78,
"\000\040\200\000\006\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000\000"},
{335, "argument", 0, 4, states_79,
"\000\040\200\000\006\000\000\000\000\000\010\000\000\000\020\002\000\300\220\050\037\000\000"},
{336, "comp_iter", 0, 2, states_80,
"\000\000\040\000\000\000\000\000\000\000\000\000\042\000\000\000\000\000\000\000\000\000\000"},
{337, "sync_comp_for", 0, 6, states_81,
"\000\000\000\000\000\000\000\000\000\000\000\000\040\000\000\000\000\000\000\000\000\000\000"},
{338, "comp_for", 0, 3, states_82,
"\000\000\040\000\000\000\000\000\000\000\000\000\040\000\000\000\000\000\000\000\000\000\000"},
{339, "comp_if", 0, 4, states_83,
"\000\000\000\000\000\000\000\000\000\000\000\000\002\000\000\000\000\000\000\000\000\000\000"},
{340, "encoding_decl", 0, 2, states_84,
"\000\000\200\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000"},
{341, "yield_expr", 0, 3, states_85,
"\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\200\000"},
{342, "yield_arg", 0, 3, states_86,
"\000\040\200\000\000\000\000\000\000\040\010\000\000\000\020\002\000\300\220\050\037\000\000"},
};
static label labels[177] = {
{0, "EMPTY"}, {0, "EMPTY"},
{256, 0}, {256, 0},
{4, 0}, {4, 0},
@ -2099,7 +2111,7 @@ static label labels[176] = {
{333, 0}, {333, 0},
{263, 0}, {263, 0},
{262, 0}, {262, 0},
{55, 0}, {1, "async"},
{1, "def"}, {1, "def"},
{1, 0}, {1, 0},
{264, 0}, {264, 0},
@ -2128,7 +2140,7 @@ static label labels[176] = {
{274, 0}, {274, 0},
{273, 0}, {273, 0},
{275, 0}, {275, 0},
{340, 0}, {341, 0},
{314, 0}, {314, 0},
{36, 0}, {36, 0},
{37, 0}, {37, 0},
@ -2229,7 +2241,7 @@ static label labels[176] = {
{31, 0}, {31, 0},
{322, 0}, {322, 0},
{323, 0}, {323, 0},
{54, 0}, {1, "await"},
{324, 0}, {324, 0},
{326, 0}, {326, 0},
{325, 0}, {325, 0},
@ -2243,21 +2255,22 @@ static label labels[176] = {
{1, "None"}, {1, "None"},
{1, "True"}, {1, "True"},
{1, "False"}, {1, "False"},
{337, 0}, {338, 0},
{327, 0}, {327, 0},
{328, 0}, {328, 0},
{329, 0}, {329, 0},
{1, "class"}, {1, "class"},
{335, 0}, {335, 0},
{336, 0}, {336, 0},
{338, 0},
{339, 0}, {339, 0},
{337, 0},
{340, 0},
{1, "yield"}, {1, "yield"},
{341, 0}, {342, 0},
}; };
grammar _PyParser_Grammar = { grammar _PyParser_Grammar = {
86, 87,
dfas, dfas,
{176, labels}, {177, labels},
256 256
}; };