From ff8d0873aabe54009af533f9f6a76fa91392a80a Mon Sep 17 00:00:00 2001 From: Berker Peksag Date: Wed, 30 Dec 2015 01:41:58 +0200 Subject: [PATCH] Issue #25977: Fix typos in Lib/tokenize.py Patch by John Walker. --- Lib/tokenize.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/Lib/tokenize.py b/Lib/tokenize.py index 65d06e53f3b..9fd676c5b24 100644 --- a/Lib/tokenize.py +++ b/Lib/tokenize.py @@ -328,8 +328,8 @@ def untokenize(iterable): Round-trip invariant for full input: Untokenized source will match input source exactly - Round-trip invariant for limited intput: - # Output bytes will tokenize the back to the input + Round-trip invariant for limited input: + # Output bytes will tokenize back to the input t1 = [tok[:2] for tok in tokenize(f.readline)] newcode = untokenize(t1) readline = BytesIO(newcode).readline @@ -465,10 +465,10 @@ def open(filename): def tokenize(readline): """ - The tokenize() generator requires one argment, readline, which + The tokenize() generator requires one argument, readline, which must be a callable object which provides the same interface as the readline() method of built-in file objects. Each call to the function - should return one line of input as bytes. Alternately, readline + should return one line of input as bytes. Alternatively, readline can be a callable function terminating with StopIteration: readline = open(myfile, 'rb').__next__ # Example of alternate readline