mirror of https://github.com/python/cpython
Tim's latest version (supports old and new tokenize modules)
This commit is contained in:
parent
5d97ebf2ac
commit
f4b44fa6ef
|
@ -1,29 +1,16 @@
|
|||
#! /home/guido/python/src/sparc/python
|
||||
#! /usr/bin/env python
|
||||
|
||||
"""The Tab Nanny despises ambiguous indentation. She knows no mercy.
|
||||
"""The Tab Nanny despises ambiguous indentation. She knows no mercy."""
|
||||
|
||||
CAUTION: this version requires Guido's "NL" patch to lib/tokenize.py,
|
||||
posted 30-Mar-98. This version will not run at all with an unpatched
|
||||
tokenize (it will raise AttributeError while loading), while previous
|
||||
versions will run incorrectly with the patched tokenize.
|
||||
"""
|
||||
# Released to the public domain, by Tim Peters, 4 April 1998.
|
||||
|
||||
# Released to the public domain, by Tim Peters, 30 March 1998.
|
||||
|
||||
__version__ = "2"
|
||||
__version__ = "3"
|
||||
|
||||
import os
|
||||
import sys
|
||||
import getopt
|
||||
import tokenize
|
||||
|
||||
try:
|
||||
tokenize.NL
|
||||
except AttributeError:
|
||||
raise AttributeError, "Sorry, I need a version of tokenize.py " \
|
||||
"that supports the NL pseudo-token."
|
||||
|
||||
verbose = 0
|
||||
|
||||
def main():
|
||||
|
@ -235,15 +222,22 @@ def format_witnesses(w):
|
|||
prefix = prefix + "s"
|
||||
return prefix + " " + string.join(firsts, ', ')
|
||||
|
||||
indents = []
|
||||
check_equal = 0
|
||||
# The collection of globals, the reset_globals() function, and the
|
||||
# tokeneater() function, depend on which version of tokenize is
|
||||
# in use.
|
||||
|
||||
def reset_globals():
|
||||
if hasattr(tokenize, 'NL'):
|
||||
# take advantage of Guido's patch!
|
||||
|
||||
indents = []
|
||||
check_equal = 0
|
||||
|
||||
def reset_globals():
|
||||
global indents, check_equal
|
||||
check_equal = 0
|
||||
indents = [Whitespace("")]
|
||||
|
||||
def tokeneater(type, token, start, end, line,
|
||||
def tokeneater(type, token, start, end, line,
|
||||
INDENT=tokenize.INDENT,
|
||||
DEDENT=tokenize.DEDENT,
|
||||
NEWLINE=tokenize.NEWLINE,
|
||||
|
@ -285,7 +279,7 @@ def tokeneater(type, token, start, end, line,
|
|||
|
||||
elif check_equal:
|
||||
# this is the first "real token" following a NEWLINE, so it
|
||||
# must be the first token of the next program statment, or an
|
||||
# must be the first token of the next program statement, or an
|
||||
# ENDMARKER; the "line" argument exposes the leading whitespace
|
||||
# for this statement; in the case of ENDMARKER, line is an empty
|
||||
# string, so will properly match the empty string with which the
|
||||
|
@ -297,6 +291,63 @@ def tokeneater(type, token, start, end, line,
|
|||
msg = "indent not equal e.g. " + format_witnesses(witness)
|
||||
raise NannyNag(start[0], msg, line)
|
||||
|
||||
else:
|
||||
# unpatched version of tokenize
|
||||
|
||||
nesting_level = 0
|
||||
indents = []
|
||||
check_equal = 0
|
||||
|
||||
def reset_globals():
|
||||
global nesting_level, indents, check_equal
|
||||
nesting_level = check_equal = 0
|
||||
indents = [Whitespace("")]
|
||||
|
||||
def tokeneater(type, token, start, end, line,
|
||||
INDENT=tokenize.INDENT,
|
||||
DEDENT=tokenize.DEDENT,
|
||||
NEWLINE=tokenize.NEWLINE,
|
||||
COMMENT=tokenize.COMMENT,
|
||||
OP=tokenize.OP):
|
||||
global nesting_level, indents, check_equal
|
||||
|
||||
if type == INDENT:
|
||||
check_equal = 0
|
||||
thisguy = Whitespace(token)
|
||||
if not indents[-1].less(thisguy):
|
||||
witness = indents[-1].not_less_witness(thisguy)
|
||||
msg = "indent not greater e.g. " + format_witnesses(witness)
|
||||
raise NannyNag(start[0], msg, line)
|
||||
indents.append(thisguy)
|
||||
|
||||
elif type == DEDENT:
|
||||
del indents[-1]
|
||||
|
||||
elif type == NEWLINE:
|
||||
if nesting_level == 0:
|
||||
check_equal = 1
|
||||
|
||||
elif type == COMMENT:
|
||||
pass
|
||||
|
||||
elif check_equal:
|
||||
check_equal = 0
|
||||
thisguy = Whitespace(line)
|
||||
if not indents[-1].equal(thisguy):
|
||||
witness = indents[-1].not_equal_witness(thisguy)
|
||||
msg = "indent not equal e.g. " + format_witnesses(witness)
|
||||
raise NannyNag(start[0], msg, line)
|
||||
|
||||
if type == OP and token in ('{', '[', '('):
|
||||
nesting_level = nesting_level + 1
|
||||
|
||||
elif type == OP and token in ('}', ']', ')'):
|
||||
if nesting_level == 0:
|
||||
raise NannyNag(start[0],
|
||||
"unbalanced bracket '" + token + "'",
|
||||
line)
|
||||
nesting_level = nesting_level - 1
|
||||
|
||||
if __name__ == '__main__':
|
||||
main()
|
||||
|
||||
|
|
|
@ -1,29 +1,16 @@
|
|||
#! /home/guido/python/src/sparc/python
|
||||
#! /usr/bin/env python
|
||||
|
||||
"""The Tab Nanny despises ambiguous indentation. She knows no mercy.
|
||||
"""The Tab Nanny despises ambiguous indentation. She knows no mercy."""
|
||||
|
||||
CAUTION: this version requires Guido's "NL" patch to lib/tokenize.py,
|
||||
posted 30-Mar-98. This version will not run at all with an unpatched
|
||||
tokenize (it will raise AttributeError while loading), while previous
|
||||
versions will run incorrectly with the patched tokenize.
|
||||
"""
|
||||
# Released to the public domain, by Tim Peters, 4 April 1998.
|
||||
|
||||
# Released to the public domain, by Tim Peters, 30 March 1998.
|
||||
|
||||
__version__ = "2"
|
||||
__version__ = "3"
|
||||
|
||||
import os
|
||||
import sys
|
||||
import getopt
|
||||
import tokenize
|
||||
|
||||
try:
|
||||
tokenize.NL
|
||||
except AttributeError:
|
||||
raise AttributeError, "Sorry, I need a version of tokenize.py " \
|
||||
"that supports the NL pseudo-token."
|
||||
|
||||
verbose = 0
|
||||
|
||||
def main():
|
||||
|
@ -235,15 +222,22 @@ def format_witnesses(w):
|
|||
prefix = prefix + "s"
|
||||
return prefix + " " + string.join(firsts, ', ')
|
||||
|
||||
indents = []
|
||||
check_equal = 0
|
||||
# The collection of globals, the reset_globals() function, and the
|
||||
# tokeneater() function, depend on which version of tokenize is
|
||||
# in use.
|
||||
|
||||
def reset_globals():
|
||||
if hasattr(tokenize, 'NL'):
|
||||
# take advantage of Guido's patch!
|
||||
|
||||
indents = []
|
||||
check_equal = 0
|
||||
|
||||
def reset_globals():
|
||||
global indents, check_equal
|
||||
check_equal = 0
|
||||
indents = [Whitespace("")]
|
||||
|
||||
def tokeneater(type, token, start, end, line,
|
||||
def tokeneater(type, token, start, end, line,
|
||||
INDENT=tokenize.INDENT,
|
||||
DEDENT=tokenize.DEDENT,
|
||||
NEWLINE=tokenize.NEWLINE,
|
||||
|
@ -285,7 +279,7 @@ def tokeneater(type, token, start, end, line,
|
|||
|
||||
elif check_equal:
|
||||
# this is the first "real token" following a NEWLINE, so it
|
||||
# must be the first token of the next program statment, or an
|
||||
# must be the first token of the next program statement, or an
|
||||
# ENDMARKER; the "line" argument exposes the leading whitespace
|
||||
# for this statement; in the case of ENDMARKER, line is an empty
|
||||
# string, so will properly match the empty string with which the
|
||||
|
@ -297,6 +291,63 @@ def tokeneater(type, token, start, end, line,
|
|||
msg = "indent not equal e.g. " + format_witnesses(witness)
|
||||
raise NannyNag(start[0], msg, line)
|
||||
|
||||
else:
|
||||
# unpatched version of tokenize
|
||||
|
||||
nesting_level = 0
|
||||
indents = []
|
||||
check_equal = 0
|
||||
|
||||
def reset_globals():
|
||||
global nesting_level, indents, check_equal
|
||||
nesting_level = check_equal = 0
|
||||
indents = [Whitespace("")]
|
||||
|
||||
def tokeneater(type, token, start, end, line,
|
||||
INDENT=tokenize.INDENT,
|
||||
DEDENT=tokenize.DEDENT,
|
||||
NEWLINE=tokenize.NEWLINE,
|
||||
COMMENT=tokenize.COMMENT,
|
||||
OP=tokenize.OP):
|
||||
global nesting_level, indents, check_equal
|
||||
|
||||
if type == INDENT:
|
||||
check_equal = 0
|
||||
thisguy = Whitespace(token)
|
||||
if not indents[-1].less(thisguy):
|
||||
witness = indents[-1].not_less_witness(thisguy)
|
||||
msg = "indent not greater e.g. " + format_witnesses(witness)
|
||||
raise NannyNag(start[0], msg, line)
|
||||
indents.append(thisguy)
|
||||
|
||||
elif type == DEDENT:
|
||||
del indents[-1]
|
||||
|
||||
elif type == NEWLINE:
|
||||
if nesting_level == 0:
|
||||
check_equal = 1
|
||||
|
||||
elif type == COMMENT:
|
||||
pass
|
||||
|
||||
elif check_equal:
|
||||
check_equal = 0
|
||||
thisguy = Whitespace(line)
|
||||
if not indents[-1].equal(thisguy):
|
||||
witness = indents[-1].not_equal_witness(thisguy)
|
||||
msg = "indent not equal e.g. " + format_witnesses(witness)
|
||||
raise NannyNag(start[0], msg, line)
|
||||
|
||||
if type == OP and token in ('{', '[', '('):
|
||||
nesting_level = nesting_level + 1
|
||||
|
||||
elif type == OP and token in ('}', ']', ')'):
|
||||
if nesting_level == 0:
|
||||
raise NannyNag(start[0],
|
||||
"unbalanced bracket '" + token + "'",
|
||||
line)
|
||||
nesting_level = nesting_level - 1
|
||||
|
||||
if __name__ == '__main__':
|
||||
main()
|
||||
|
||||
|
|
|
@ -1,29 +1,16 @@
|
|||
#! /home/guido/python/src/sparc/python
|
||||
#! /usr/bin/env python
|
||||
|
||||
"""The Tab Nanny despises ambiguous indentation. She knows no mercy.
|
||||
"""The Tab Nanny despises ambiguous indentation. She knows no mercy."""
|
||||
|
||||
CAUTION: this version requires Guido's "NL" patch to lib/tokenize.py,
|
||||
posted 30-Mar-98. This version will not run at all with an unpatched
|
||||
tokenize (it will raise AttributeError while loading), while previous
|
||||
versions will run incorrectly with the patched tokenize.
|
||||
"""
|
||||
# Released to the public domain, by Tim Peters, 4 April 1998.
|
||||
|
||||
# Released to the public domain, by Tim Peters, 30 March 1998.
|
||||
|
||||
__version__ = "2"
|
||||
__version__ = "3"
|
||||
|
||||
import os
|
||||
import sys
|
||||
import getopt
|
||||
import tokenize
|
||||
|
||||
try:
|
||||
tokenize.NL
|
||||
except AttributeError:
|
||||
raise AttributeError, "Sorry, I need a version of tokenize.py " \
|
||||
"that supports the NL pseudo-token."
|
||||
|
||||
verbose = 0
|
||||
|
||||
def main():
|
||||
|
@ -235,15 +222,22 @@ def format_witnesses(w):
|
|||
prefix = prefix + "s"
|
||||
return prefix + " " + string.join(firsts, ', ')
|
||||
|
||||
indents = []
|
||||
check_equal = 0
|
||||
# The collection of globals, the reset_globals() function, and the
|
||||
# tokeneater() function, depend on which version of tokenize is
|
||||
# in use.
|
||||
|
||||
def reset_globals():
|
||||
if hasattr(tokenize, 'NL'):
|
||||
# take advantage of Guido's patch!
|
||||
|
||||
indents = []
|
||||
check_equal = 0
|
||||
|
||||
def reset_globals():
|
||||
global indents, check_equal
|
||||
check_equal = 0
|
||||
indents = [Whitespace("")]
|
||||
|
||||
def tokeneater(type, token, start, end, line,
|
||||
def tokeneater(type, token, start, end, line,
|
||||
INDENT=tokenize.INDENT,
|
||||
DEDENT=tokenize.DEDENT,
|
||||
NEWLINE=tokenize.NEWLINE,
|
||||
|
@ -285,7 +279,7 @@ def tokeneater(type, token, start, end, line,
|
|||
|
||||
elif check_equal:
|
||||
# this is the first "real token" following a NEWLINE, so it
|
||||
# must be the first token of the next program statment, or an
|
||||
# must be the first token of the next program statement, or an
|
||||
# ENDMARKER; the "line" argument exposes the leading whitespace
|
||||
# for this statement; in the case of ENDMARKER, line is an empty
|
||||
# string, so will properly match the empty string with which the
|
||||
|
@ -297,6 +291,63 @@ def tokeneater(type, token, start, end, line,
|
|||
msg = "indent not equal e.g. " + format_witnesses(witness)
|
||||
raise NannyNag(start[0], msg, line)
|
||||
|
||||
else:
|
||||
# unpatched version of tokenize
|
||||
|
||||
nesting_level = 0
|
||||
indents = []
|
||||
check_equal = 0
|
||||
|
||||
def reset_globals():
|
||||
global nesting_level, indents, check_equal
|
||||
nesting_level = check_equal = 0
|
||||
indents = [Whitespace("")]
|
||||
|
||||
def tokeneater(type, token, start, end, line,
|
||||
INDENT=tokenize.INDENT,
|
||||
DEDENT=tokenize.DEDENT,
|
||||
NEWLINE=tokenize.NEWLINE,
|
||||
COMMENT=tokenize.COMMENT,
|
||||
OP=tokenize.OP):
|
||||
global nesting_level, indents, check_equal
|
||||
|
||||
if type == INDENT:
|
||||
check_equal = 0
|
||||
thisguy = Whitespace(token)
|
||||
if not indents[-1].less(thisguy):
|
||||
witness = indents[-1].not_less_witness(thisguy)
|
||||
msg = "indent not greater e.g. " + format_witnesses(witness)
|
||||
raise NannyNag(start[0], msg, line)
|
||||
indents.append(thisguy)
|
||||
|
||||
elif type == DEDENT:
|
||||
del indents[-1]
|
||||
|
||||
elif type == NEWLINE:
|
||||
if nesting_level == 0:
|
||||
check_equal = 1
|
||||
|
||||
elif type == COMMENT:
|
||||
pass
|
||||
|
||||
elif check_equal:
|
||||
check_equal = 0
|
||||
thisguy = Whitespace(line)
|
||||
if not indents[-1].equal(thisguy):
|
||||
witness = indents[-1].not_equal_witness(thisguy)
|
||||
msg = "indent not equal e.g. " + format_witnesses(witness)
|
||||
raise NannyNag(start[0], msg, line)
|
||||
|
||||
if type == OP and token in ('{', '[', '('):
|
||||
nesting_level = nesting_level + 1
|
||||
|
||||
elif type == OP and token in ('}', ']', ')'):
|
||||
if nesting_level == 0:
|
||||
raise NannyNag(start[0],
|
||||
"unbalanced bracket '" + token + "'",
|
||||
line)
|
||||
nesting_level = nesting_level - 1
|
||||
|
||||
if __name__ == '__main__':
|
||||
main()
|
||||
|
||||
|
|
Loading…
Reference in New Issue