mirror of https://github.com/python/cpython
d87b015106
* gh-119118: Fix performance regression in tokenize module - Cache line object to avoid creating a Unicode object for all of the tokens in the same line. - Speed up byte offset to column offset conversion by using the smallest buffer possible to measure the difference. Co-authored-by: Pablo Galindo <pablogsal@gmail.com> |
||
---|---|---|
.. | ||
lexer | ||
tokenizer | ||
Python.asdl | ||
action_helpers.c | ||
asdl.py | ||
asdl_c.py | ||
myreadline.c | ||
parser.c | ||
peg_api.c | ||
pegen.c | ||
pegen.h | ||
pegen_errors.c | ||
string_parser.c | ||
string_parser.h | ||
token.c |