2018-01-20 07:23:59 -04:00
|
|
|
"""
|
|
|
|
The typing module: Support for gradual typing as defined by PEP 484.
|
|
|
|
|
|
|
|
At large scale, the structure of the module is following:
|
2018-09-01 05:56:58 -03:00
|
|
|
* Imports and exports, all public names should be explicitly added to __all__.
|
2018-01-20 07:23:59 -04:00
|
|
|
* Internal helper functions: these should never be used in code outside this module.
|
|
|
|
* _SpecialForm and its instances (special forms): Any, NoReturn, ClassVar, Union, Optional
|
|
|
|
* Two classes whose instances can be type arguments in addition to types: ForwardRef and TypeVar
|
|
|
|
* The core of internal generics API: _GenericAlias and _VariadicGenericAlias, the latter is
|
|
|
|
currently only used by Tuple and Callable. All subscripted types like X[int], Union[int, str],
|
|
|
|
etc., are instances of either of these classes.
|
|
|
|
* The public counterpart of the generics API consists of two classes: Generic and Protocol
|
|
|
|
(the latter is currently private, but will be made public after PEP 544 acceptance).
|
|
|
|
* Public helper functions: get_type_hints, overload, cast, no_type_check,
|
|
|
|
no_type_check_decorator.
|
|
|
|
* Generic aliases for collections.abc ABCs and few additional protocols.
|
|
|
|
* Special types: NewType, NamedTuple, TypedDict (may be added soon).
|
|
|
|
* Wrapper submodules for re and io related types.
|
|
|
|
"""
|
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
from abc import abstractmethod, abstractproperty
|
|
|
|
import collections
|
2018-01-20 07:23:59 -04:00
|
|
|
import collections.abc
|
2016-04-15 14:51:30 -03:00
|
|
|
import contextlib
|
2015-05-22 14:14:11 -03:00
|
|
|
import functools
|
2018-05-26 15:19:26 -03:00
|
|
|
import operator
|
2015-05-22 14:14:11 -03:00
|
|
|
import re as stdlib_re # Avoid confusion with the re we export.
|
|
|
|
import sys
|
|
|
|
import types
|
2018-01-20 07:23:59 -04:00
|
|
|
from types import WrapperDescriptorType, MethodWrapperType, MethodDescriptorType
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
# Please keep __all__ alphabetized within each category.
|
|
|
|
__all__ = [
|
|
|
|
# Super-special typing primitives.
|
|
|
|
'Any',
|
|
|
|
'Callable',
|
2016-09-11 19:34:56 -03:00
|
|
|
'ClassVar',
|
2015-05-22 14:14:11 -03:00
|
|
|
'Generic',
|
|
|
|
'Optional',
|
2016-05-24 20:38:22 -03:00
|
|
|
'Tuple',
|
|
|
|
'Type',
|
2015-05-22 14:14:11 -03:00
|
|
|
'TypeVar',
|
|
|
|
'Union',
|
|
|
|
|
|
|
|
# ABCs (from collections.abc).
|
|
|
|
'AbstractSet', # collections.abc.Set.
|
|
|
|
'ByteString',
|
|
|
|
'Container',
|
2017-06-10 16:57:56 -03:00
|
|
|
'ContextManager',
|
2015-05-22 14:14:11 -03:00
|
|
|
'Hashable',
|
|
|
|
'ItemsView',
|
|
|
|
'Iterable',
|
|
|
|
'Iterator',
|
|
|
|
'KeysView',
|
|
|
|
'Mapping',
|
|
|
|
'MappingView',
|
|
|
|
'MutableMapping',
|
|
|
|
'MutableSequence',
|
|
|
|
'MutableSet',
|
|
|
|
'Sequence',
|
|
|
|
'Sized',
|
|
|
|
'ValuesView',
|
2018-01-20 07:23:59 -04:00
|
|
|
'Awaitable',
|
|
|
|
'AsyncIterator',
|
|
|
|
'AsyncIterable',
|
|
|
|
'Coroutine',
|
|
|
|
'Collection',
|
|
|
|
'AsyncGenerator',
|
|
|
|
'AsyncContextManager',
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
# Structural checks, a.k.a. protocols.
|
|
|
|
'Reversible',
|
|
|
|
'SupportsAbs',
|
2017-05-02 14:14:07 -03:00
|
|
|
'SupportsBytes',
|
|
|
|
'SupportsComplex',
|
2015-05-22 14:14:11 -03:00
|
|
|
'SupportsFloat',
|
|
|
|
'SupportsInt',
|
|
|
|
'SupportsRound',
|
|
|
|
|
|
|
|
# Concrete collection types.
|
2017-02-13 17:50:14 -04:00
|
|
|
'Counter',
|
2017-01-17 02:42:37 -04:00
|
|
|
'Deque',
|
2015-05-22 14:14:11 -03:00
|
|
|
'Dict',
|
2016-04-05 12:28:52 -03:00
|
|
|
'DefaultDict',
|
2015-05-22 14:14:11 -03:00
|
|
|
'List',
|
|
|
|
'Set',
|
2016-08-23 15:01:50 -03:00
|
|
|
'FrozenSet',
|
2015-05-22 14:14:11 -03:00
|
|
|
'NamedTuple', # Not really a type.
|
|
|
|
'Generator',
|
|
|
|
|
|
|
|
# One-off things.
|
|
|
|
'AnyStr',
|
|
|
|
'cast',
|
|
|
|
'get_type_hints',
|
2016-06-08 15:19:11 -03:00
|
|
|
'NewType',
|
2015-05-22 14:14:11 -03:00
|
|
|
'no_type_check',
|
|
|
|
'no_type_check_decorator',
|
2018-03-19 15:41:32 -03:00
|
|
|
'NoReturn',
|
2015-05-22 14:14:11 -03:00
|
|
|
'overload',
|
2016-04-05 18:54:25 -03:00
|
|
|
'Text',
|
2016-06-08 15:19:11 -03:00
|
|
|
'TYPE_CHECKING',
|
2015-05-22 14:14:11 -03:00
|
|
|
]
|
|
|
|
|
2016-04-05 12:28:52 -03:00
|
|
|
# The pseudo-submodules 're' and 'io' are part of the public
|
|
|
|
# namespace, but excluded from __all__ because they might stomp on
|
|
|
|
# legitimate imports of those modules.
|
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-05-23 00:32:10 -03:00
|
|
|
def _type_check(arg, msg, is_argument=True):
|
2018-01-20 07:23:59 -04:00
|
|
|
"""Check that the argument is a type, and return it (internal helper).
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
As a special case, accept None and return type(None) instead. Also wrap strings
|
|
|
|
into ForwardRef instances. Consider several corner cases, for example plain
|
|
|
|
special forms like Union are not valid, while Union[int, str] is OK, etc.
|
|
|
|
The msg argument is a human-readable error message, e.g::
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
"Union[arg, ...]: arg should be a type."
|
2016-09-27 19:20:12 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
We append the repr() of the actual value (truncated to 100 chars).
|
|
|
|
"""
|
2018-05-16 13:27:03 -03:00
|
|
|
invalid_generic_forms = (Generic, _Protocol)
|
2018-05-23 00:32:10 -03:00
|
|
|
if is_argument:
|
2018-05-16 13:27:03 -03:00
|
|
|
invalid_generic_forms = invalid_generic_forms + (ClassVar, )
|
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
if arg is None:
|
|
|
|
return type(None)
|
|
|
|
if isinstance(arg, str):
|
|
|
|
return ForwardRef(arg)
|
|
|
|
if (isinstance(arg, _GenericAlias) and
|
2018-05-16 13:27:03 -03:00
|
|
|
arg.__origin__ in invalid_generic_forms):
|
2018-01-20 07:23:59 -04:00
|
|
|
raise TypeError(f"{arg} is not valid as type argument")
|
2018-10-08 15:50:16 -03:00
|
|
|
if (isinstance(arg, _SpecialForm) and arg not in (Any, NoReturn) or
|
2018-01-20 07:23:59 -04:00
|
|
|
arg in (Generic, _Protocol)):
|
|
|
|
raise TypeError(f"Plain {arg} is not valid as type argument")
|
|
|
|
if isinstance(arg, (type, TypeVar, ForwardRef)):
|
|
|
|
return arg
|
|
|
|
if not callable(arg):
|
|
|
|
raise TypeError(f"{msg} Got {arg!r:.100}.")
|
|
|
|
return arg
|
2016-09-27 19:20:12 -03:00
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def _type_repr(obj):
|
|
|
|
"""Return the repr() of an object, special-casing types (internal helper).
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
If obj is a type, we return a shorter version than the default
|
|
|
|
type.__repr__, based on the module and qualified name, which is
|
|
|
|
typically enough to uniquely identify a type. For everything
|
|
|
|
else, we fall back on repr(obj).
|
2015-05-22 14:14:11 -03:00
|
|
|
"""
|
2018-01-20 07:23:59 -04:00
|
|
|
if isinstance(obj, type):
|
|
|
|
if obj.__module__ == 'builtins':
|
|
|
|
return obj.__qualname__
|
|
|
|
return f'{obj.__module__}.{obj.__qualname__}'
|
|
|
|
if obj is ...:
|
|
|
|
return('...')
|
|
|
|
if isinstance(obj, types.FunctionType):
|
|
|
|
return obj.__name__
|
|
|
|
return repr(obj)
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def _collect_type_vars(types):
|
|
|
|
"""Collect all type variable contained in types in order of
|
|
|
|
first appearance (lexicographic order). For example::
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
_collect_type_vars((T, List[S, T])) == (T, S)
|
|
|
|
"""
|
|
|
|
tvars = []
|
|
|
|
for t in types:
|
|
|
|
if isinstance(t, TypeVar) and t not in tvars:
|
|
|
|
tvars.append(t)
|
|
|
|
if isinstance(t, _GenericAlias) and not t._special:
|
|
|
|
tvars.extend([t for t in t.__parameters__ if t not in tvars])
|
|
|
|
return tuple(tvars)
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def _subs_tvars(tp, tvars, subs):
|
|
|
|
"""Substitute type variables 'tvars' with substitutions 'subs'.
|
|
|
|
These two must have the same length.
|
|
|
|
"""
|
|
|
|
if not isinstance(tp, _GenericAlias):
|
|
|
|
return tp
|
|
|
|
new_args = list(tp.__args__)
|
|
|
|
for a, arg in enumerate(tp.__args__):
|
|
|
|
if isinstance(arg, TypeVar):
|
|
|
|
for i, tvar in enumerate(tvars):
|
|
|
|
if arg == tvar:
|
|
|
|
new_args[a] = subs[i]
|
|
|
|
else:
|
|
|
|
new_args[a] = _subs_tvars(arg, tvars, subs)
|
|
|
|
if tp.__origin__ is Union:
|
|
|
|
return Union[tuple(new_args)]
|
|
|
|
return tp.copy_with(tuple(new_args))
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def _check_generic(cls, parameters):
|
|
|
|
"""Check correct count for parameters of a generic cls (internal helper).
|
|
|
|
This gives a nice error message in case of count mismatch.
|
|
|
|
"""
|
|
|
|
if not cls.__parameters__:
|
|
|
|
raise TypeError(f"{cls} is not a generic class")
|
|
|
|
alen = len(parameters)
|
|
|
|
elen = len(cls.__parameters__)
|
|
|
|
if alen != elen:
|
|
|
|
raise TypeError(f"Too {'many' if alen > elen else 'few'} parameters for {cls};"
|
|
|
|
f" actual {alen}, expected {elen}")
|
|
|
|
|
|
|
|
|
|
|
|
def _remove_dups_flatten(parameters):
|
2018-05-18 20:00:38 -03:00
|
|
|
"""An internal helper for Union creation and substitution: flatten Unions
|
|
|
|
among parameters, then remove duplicates.
|
2018-01-20 07:23:59 -04:00
|
|
|
"""
|
|
|
|
# Flatten out Union[Union[...], ...].
|
|
|
|
params = []
|
|
|
|
for p in parameters:
|
|
|
|
if isinstance(p, _GenericAlias) and p.__origin__ is Union:
|
|
|
|
params.extend(p.__args__)
|
|
|
|
elif isinstance(p, tuple) and len(p) > 0 and p[0] is Union:
|
|
|
|
params.extend(p[1:])
|
|
|
|
else:
|
|
|
|
params.append(p)
|
|
|
|
# Weed out strict duplicates, preserving the first of each occurrence.
|
|
|
|
all_params = set(params)
|
|
|
|
if len(all_params) < len(params):
|
|
|
|
new_params = []
|
|
|
|
for t in params:
|
|
|
|
if t in all_params:
|
|
|
|
new_params.append(t)
|
|
|
|
all_params.remove(t)
|
|
|
|
params = new_params
|
|
|
|
assert not all_params, all_params
|
2018-05-18 20:00:38 -03:00
|
|
|
return tuple(params)
|
2018-01-20 07:23:59 -04:00
|
|
|
|
|
|
|
|
|
|
|
_cleanups = []
|
|
|
|
|
|
|
|
|
|
|
|
def _tp_cache(func):
|
|
|
|
"""Internal wrapper caching __getitem__ of generic types with a fallback to
|
|
|
|
original function for non-hashable arguments.
|
|
|
|
"""
|
|
|
|
cached = functools.lru_cache()(func)
|
|
|
|
_cleanups.append(cached.cache_clear)
|
|
|
|
|
|
|
|
@functools.wraps(func)
|
|
|
|
def inner(*args, **kwds):
|
|
|
|
try:
|
|
|
|
return cached(*args, **kwds)
|
|
|
|
except TypeError:
|
|
|
|
pass # All real errors (not unhashable args) are raised below.
|
|
|
|
return func(*args, **kwds)
|
|
|
|
return inner
|
|
|
|
|
|
|
|
|
|
|
|
def _eval_type(t, globalns, localns):
|
|
|
|
"""Evaluate all forward reverences in the given type t.
|
|
|
|
For use of globalns and localns see the docstring for get_type_hints().
|
|
|
|
"""
|
|
|
|
if isinstance(t, ForwardRef):
|
|
|
|
return t._evaluate(globalns, localns)
|
|
|
|
if isinstance(t, _GenericAlias):
|
|
|
|
ev_args = tuple(_eval_type(a, globalns, localns) for a in t.__args__)
|
|
|
|
if ev_args == t.__args__:
|
|
|
|
return t
|
|
|
|
res = t.copy_with(ev_args)
|
|
|
|
res._special = t._special
|
|
|
|
return res
|
|
|
|
return t
|
2016-09-27 19:20:12 -03:00
|
|
|
|
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
class _Final:
|
|
|
|
"""Mixin to prohibit subclassing"""
|
2016-09-27 19:20:12 -03:00
|
|
|
|
2017-01-18 00:43:28 -04:00
|
|
|
__slots__ = ('__weakref__',)
|
2016-09-27 19:20:12 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def __init_subclass__(self, *args, **kwds):
|
|
|
|
if '_root' not in kwds:
|
|
|
|
raise TypeError("Cannot subclass special typing classes")
|
|
|
|
|
2018-03-26 19:01:12 -03:00
|
|
|
class _Immutable:
|
|
|
|
"""Mixin to indicate that object should not be copied."""
|
2018-01-20 07:23:59 -04:00
|
|
|
|
2018-03-26 19:01:12 -03:00
|
|
|
def __copy__(self):
|
|
|
|
return self
|
|
|
|
|
|
|
|
def __deepcopy__(self, memo):
|
|
|
|
return self
|
|
|
|
|
|
|
|
|
|
|
|
class _SpecialForm(_Final, _Immutable, _root=True):
|
2018-01-20 07:23:59 -04:00
|
|
|
"""Internal indicator of special typing constructs.
|
|
|
|
See _doc instance attribute for specific docs.
|
|
|
|
"""
|
|
|
|
|
|
|
|
__slots__ = ('_name', '_doc')
|
|
|
|
|
2016-09-27 19:20:12 -03:00
|
|
|
def __new__(cls, *args, **kwds):
|
|
|
|
"""Constructor.
|
|
|
|
|
|
|
|
This only exists to give a better error message in case
|
|
|
|
someone tries to subclass a special typing object (not a good idea).
|
|
|
|
"""
|
|
|
|
if (len(args) == 3 and
|
|
|
|
isinstance(args[0], str) and
|
|
|
|
isinstance(args[1], tuple)):
|
|
|
|
# Close enough.
|
2018-01-20 07:23:59 -04:00
|
|
|
raise TypeError(f"Cannot subclass {cls!r}")
|
2016-10-03 12:40:50 -03:00
|
|
|
return super().__new__(cls)
|
2016-09-27 19:20:12 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def __init__(self, name, doc):
|
|
|
|
self._name = name
|
|
|
|
self._doc = doc
|
2016-09-27 19:20:12 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def __eq__(self, other):
|
|
|
|
if not isinstance(other, _SpecialForm):
|
|
|
|
return NotImplemented
|
|
|
|
return self._name == other._name
|
|
|
|
|
|
|
|
def __hash__(self):
|
|
|
|
return hash((self._name,))
|
2016-09-27 19:20:12 -03:00
|
|
|
|
|
|
|
def __repr__(self):
|
2018-01-20 07:23:59 -04:00
|
|
|
return 'typing.' + self._name
|
|
|
|
|
2018-03-26 19:01:12 -03:00
|
|
|
def __reduce__(self):
|
|
|
|
return self._name
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2016-09-27 19:20:12 -03:00
|
|
|
def __call__(self, *args, **kwds):
|
2018-01-20 07:23:59 -04:00
|
|
|
raise TypeError(f"Cannot instantiate {self!r}")
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def __instancecheck__(self, obj):
|
|
|
|
raise TypeError(f"{self} cannot be used with isinstance()")
|
2016-09-27 19:20:12 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def __subclasscheck__(self, cls):
|
|
|
|
raise TypeError(f"{self} cannot be used with issubclass()")
|
2016-10-03 12:40:50 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
@_tp_cache
|
|
|
|
def __getitem__(self, parameters):
|
|
|
|
if self._name == 'ClassVar':
|
|
|
|
item = _type_check(parameters, 'ClassVar accepts only single type.')
|
|
|
|
return _GenericAlias(self, (item,))
|
|
|
|
if self._name == 'Union':
|
|
|
|
if parameters == ():
|
|
|
|
raise TypeError("Cannot take a Union of no types.")
|
|
|
|
if not isinstance(parameters, tuple):
|
|
|
|
parameters = (parameters,)
|
|
|
|
msg = "Union[arg, ...]: each arg must be a type."
|
|
|
|
parameters = tuple(_type_check(p, msg) for p in parameters)
|
|
|
|
parameters = _remove_dups_flatten(parameters)
|
|
|
|
if len(parameters) == 1:
|
|
|
|
return parameters[0]
|
|
|
|
return _GenericAlias(self, parameters)
|
|
|
|
if self._name == 'Optional':
|
|
|
|
arg = _type_check(parameters, "Optional[t] requires a single type.")
|
|
|
|
return Union[arg, type(None)]
|
|
|
|
raise TypeError(f"{self} is not subscriptable")
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2015-08-05 07:11:06 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
Any = _SpecialForm('Any', doc=
|
|
|
|
"""Special type indicating an unconstrained type.
|
|
|
|
|
|
|
|
- Any is compatible with every type.
|
|
|
|
- Any assumed to have all methods.
|
|
|
|
- All values assumed to be instances of Any.
|
|
|
|
|
|
|
|
Note that all the above statements are true from the point of view of
|
|
|
|
static type checkers. At runtime, Any should not be used with instance
|
|
|
|
or class checks.
|
|
|
|
""")
|
|
|
|
|
|
|
|
NoReturn = _SpecialForm('NoReturn', doc=
|
|
|
|
"""Special type indicating functions that never return.
|
|
|
|
Example::
|
|
|
|
|
|
|
|
from typing import NoReturn
|
|
|
|
|
|
|
|
def stop() -> NoReturn:
|
|
|
|
raise Exception('no way')
|
|
|
|
|
|
|
|
This type is invalid in other positions, e.g., ``List[NoReturn]``
|
|
|
|
will fail in static type checkers.
|
|
|
|
""")
|
|
|
|
|
|
|
|
ClassVar = _SpecialForm('ClassVar', doc=
|
|
|
|
"""Special type construct to mark class variables.
|
|
|
|
|
|
|
|
An annotation wrapped in ClassVar indicates that a given
|
|
|
|
attribute is intended to be used as a class variable and
|
|
|
|
should not be set on instances of that class. Usage::
|
|
|
|
|
|
|
|
class Starship:
|
|
|
|
stats: ClassVar[Dict[str, int]] = {} # class variable
|
|
|
|
damage: int = 10 # instance variable
|
|
|
|
|
|
|
|
ClassVar accepts only types and cannot be further subscribed.
|
|
|
|
|
|
|
|
Note that ClassVar is not a class itself, and should not
|
|
|
|
be used with isinstance() or issubclass().
|
|
|
|
""")
|
|
|
|
|
|
|
|
Union = _SpecialForm('Union', doc=
|
|
|
|
"""Union type; Union[X, Y] means either X or Y.
|
|
|
|
|
|
|
|
To define a union, use e.g. Union[int, str]. Details:
|
|
|
|
- The arguments must be types and there must be at least one.
|
|
|
|
- None as an argument is a special case and is replaced by
|
|
|
|
type(None).
|
|
|
|
- Unions of unions are flattened, e.g.::
|
|
|
|
|
|
|
|
Union[Union[int, str], float] == Union[int, str, float]
|
|
|
|
|
|
|
|
- Unions of a single argument vanish, e.g.::
|
|
|
|
|
|
|
|
Union[int] == int # The constructor actually returns int
|
|
|
|
|
|
|
|
- Redundant arguments are skipped, e.g.::
|
|
|
|
|
|
|
|
Union[int, str, int] == Union[int, str]
|
|
|
|
|
|
|
|
- When comparing unions, the argument order is ignored, e.g.::
|
|
|
|
|
|
|
|
Union[int, str] == Union[str, int]
|
|
|
|
|
|
|
|
- You cannot subclass or instantiate a union.
|
|
|
|
- You can use Optional[X] as a shorthand for Union[X, None].
|
|
|
|
""")
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
Optional = _SpecialForm('Optional', doc=
|
|
|
|
"""Optional type.
|
|
|
|
|
|
|
|
Optional[X] is equivalent to Union[X, None].
|
|
|
|
""")
|
2016-10-29 16:44:29 -03:00
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
class ForwardRef(_Final, _root=True):
|
2016-11-20 22:01:29 -04:00
|
|
|
"""Internal wrapper to hold a forward reference."""
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2016-09-27 19:20:12 -03:00
|
|
|
__slots__ = ('__forward_arg__', '__forward_code__',
|
2018-05-16 13:27:03 -03:00
|
|
|
'__forward_evaluated__', '__forward_value__',
|
|
|
|
'__forward_is_argument__')
|
2016-09-27 19:20:12 -03:00
|
|
|
|
2018-05-23 00:32:10 -03:00
|
|
|
def __init__(self, arg, is_argument=True):
|
2015-05-22 14:14:11 -03:00
|
|
|
if not isinstance(arg, str):
|
2018-01-20 07:23:59 -04:00
|
|
|
raise TypeError(f"Forward reference must be a string -- got {arg!r}")
|
2015-05-22 14:14:11 -03:00
|
|
|
try:
|
|
|
|
code = compile(arg, '<string>', 'eval')
|
|
|
|
except SyntaxError:
|
2018-01-20 07:23:59 -04:00
|
|
|
raise SyntaxError(f"Forward reference must be an expression -- got {arg!r}")
|
2015-05-22 14:14:11 -03:00
|
|
|
self.__forward_arg__ = arg
|
|
|
|
self.__forward_code__ = code
|
|
|
|
self.__forward_evaluated__ = False
|
|
|
|
self.__forward_value__ = None
|
2018-05-16 13:27:03 -03:00
|
|
|
self.__forward_is_argument__ = is_argument
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def _evaluate(self, globalns, localns):
|
2016-11-10 12:29:18 -04:00
|
|
|
if not self.__forward_evaluated__ or localns is not globalns:
|
2015-05-22 14:14:11 -03:00
|
|
|
if globalns is None and localns is None:
|
|
|
|
globalns = localns = {}
|
|
|
|
elif globalns is None:
|
|
|
|
globalns = localns
|
|
|
|
elif localns is None:
|
|
|
|
localns = globalns
|
|
|
|
self.__forward_value__ = _type_check(
|
|
|
|
eval(self.__forward_code__, globalns, localns),
|
2018-05-16 13:27:03 -03:00
|
|
|
"Forward references must evaluate to types.",
|
|
|
|
is_argument=self.__forward_is_argument__)
|
2015-05-22 14:14:11 -03:00
|
|
|
self.__forward_evaluated__ = True
|
|
|
|
return self.__forward_value__
|
|
|
|
|
2016-09-27 19:20:12 -03:00
|
|
|
def __eq__(self, other):
|
2018-01-20 07:23:59 -04:00
|
|
|
if not isinstance(other, ForwardRef):
|
2016-09-27 19:20:12 -03:00
|
|
|
return NotImplemented
|
|
|
|
return (self.__forward_arg__ == other.__forward_arg__ and
|
2016-11-10 12:24:06 -04:00
|
|
|
self.__forward_value__ == other.__forward_value__)
|
2016-09-27 19:20:12 -03:00
|
|
|
|
|
|
|
def __hash__(self):
|
2016-11-10 12:24:06 -04:00
|
|
|
return hash((self.__forward_arg__, self.__forward_value__))
|
2016-09-27 19:20:12 -03:00
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
def __repr__(self):
|
2018-01-20 07:23:59 -04:00
|
|
|
return f'ForwardRef({self.__forward_arg__!r})'
|
2015-08-05 07:11:06 -03:00
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-03-26 19:01:12 -03:00
|
|
|
class TypeVar(_Final, _Immutable, _root=True):
|
2018-01-20 07:23:59 -04:00
|
|
|
"""Type variable.
|
2016-09-27 19:20:12 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
Usage::
|
2016-09-27 19:20:12 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
T = TypeVar('T') # Can be anything
|
|
|
|
A = TypeVar('A', str, bytes) # Must be str or bytes
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
Type variables exist primarily for the benefit of static type
|
|
|
|
checkers. They serve as the parameters for generic types as well
|
|
|
|
as for generic function definitions. See class Generic for more
|
|
|
|
information on generic types. Generic functions work as follows:
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def repeat(x: T, n: int) -> List[T]:
|
|
|
|
'''Return a list containing n references to x.'''
|
|
|
|
return [x]*n
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def longest(x: A, y: A) -> A:
|
|
|
|
'''Return the longest of two strings.'''
|
|
|
|
return x if len(x) >= len(y) else y
|
2016-04-05 12:28:52 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
The latter example's signature is essentially the overloading
|
|
|
|
of (str, str) -> str and (bytes, bytes) -> bytes. Also note
|
|
|
|
that if the arguments are instances of some subclass of str,
|
|
|
|
the return type is still plain str.
|
2016-04-05 12:28:52 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
At runtime, isinstance(x, T) and issubclass(C, T) will raise TypeError.
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
Type variables defined with covariant=True or contravariant=True
|
2018-07-07 12:41:20 -03:00
|
|
|
can be used to declare covariant or contravariant generic types.
|
2018-01-20 07:23:59 -04:00
|
|
|
See PEP 484 for more details. By default generic types are invariant
|
|
|
|
in all type variables.
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
Type variables can be introspected. e.g.:
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
T.__name__ == 'T'
|
|
|
|
T.__constraints__ == ()
|
|
|
|
T.__covariant__ == False
|
|
|
|
T.__contravariant__ = False
|
|
|
|
A.__constraints__ == (str, bytes)
|
2018-03-26 19:01:12 -03:00
|
|
|
|
|
|
|
Note that only type variables defined in global scope can be pickled.
|
2018-01-20 07:23:59 -04:00
|
|
|
"""
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
__slots__ = ('__name__', '__bound__', '__constraints__',
|
2018-05-26 15:19:26 -03:00
|
|
|
'__covariant__', '__contravariant__')
|
2016-09-27 19:20:12 -03:00
|
|
|
|
|
|
|
def __init__(self, name, *constraints, bound=None,
|
2017-01-22 21:43:53 -04:00
|
|
|
covariant=False, contravariant=False):
|
2016-09-27 19:20:12 -03:00
|
|
|
self.__name__ = name
|
2015-05-22 14:14:11 -03:00
|
|
|
if covariant and contravariant:
|
2016-08-23 15:01:50 -03:00
|
|
|
raise ValueError("Bivariant types are not supported.")
|
2015-05-22 14:14:11 -03:00
|
|
|
self.__covariant__ = bool(covariant)
|
|
|
|
self.__contravariant__ = bool(contravariant)
|
|
|
|
if constraints and bound is not None:
|
|
|
|
raise TypeError("Constraints cannot be combined with bound=...")
|
|
|
|
if constraints and len(constraints) == 1:
|
|
|
|
raise TypeError("A single constraint is not allowed")
|
|
|
|
msg = "TypeVar(name, constraint, ...): constraints must be types."
|
|
|
|
self.__constraints__ = tuple(_type_check(t, msg) for t in constraints)
|
|
|
|
if bound:
|
|
|
|
self.__bound__ = _type_check(bound, "Bound must be a type.")
|
|
|
|
else:
|
|
|
|
self.__bound__ = None
|
2018-05-26 15:19:26 -03:00
|
|
|
def_mod = sys._getframe(1).f_globals['__name__'] # for pickling
|
|
|
|
if def_mod != 'typing':
|
|
|
|
self.__module__ = def_mod
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
def __repr__(self):
|
|
|
|
if self.__covariant__:
|
|
|
|
prefix = '+'
|
|
|
|
elif self.__contravariant__:
|
|
|
|
prefix = '-'
|
|
|
|
else:
|
|
|
|
prefix = '~'
|
|
|
|
return prefix + self.__name__
|
|
|
|
|
2018-03-26 19:01:12 -03:00
|
|
|
def __reduce__(self):
|
2018-05-26 15:19:26 -03:00
|
|
|
return self.__name__
|
2018-03-26 19:01:12 -03:00
|
|
|
|
2016-10-29 12:54:56 -03:00
|
|
|
|
2017-01-18 00:43:28 -04:00
|
|
|
# Special typing constructs Union, Optional, Generic, Callable and Tuple
|
|
|
|
# use three special attributes for internal bookkeeping of generic types:
|
|
|
|
# * __parameters__ is a tuple of unique free type parameters of a generic
|
|
|
|
# type, for example, Dict[T, T].__parameters__ == (T,);
|
|
|
|
# * __origin__ keeps a reference to a type that was subscripted,
|
2018-05-08 22:23:46 -03:00
|
|
|
# e.g., Union[T, int].__origin__ == Union, or the non-generic version of
|
|
|
|
# the type.
|
2017-01-18 00:43:28 -04:00
|
|
|
# * __args__ is a tuple of all arguments used in subscripting,
|
|
|
|
# e.g., Dict[T, int].__args__ == (T, int).
|
|
|
|
|
2018-04-04 21:25:15 -03:00
|
|
|
|
|
|
|
# Mapping from non-generic type names that have a generic alias in typing
|
|
|
|
# but with a different name.
|
|
|
|
_normalize_alias = {'list': 'List',
|
|
|
|
'tuple': 'Tuple',
|
|
|
|
'dict': 'Dict',
|
|
|
|
'set': 'Set',
|
|
|
|
'frozenset': 'FrozenSet',
|
|
|
|
'deque': 'Deque',
|
|
|
|
'defaultdict': 'DefaultDict',
|
|
|
|
'type': 'Type',
|
|
|
|
'Set': 'AbstractSet'}
|
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def _is_dunder(attr):
|
|
|
|
return attr.startswith('__') and attr.endswith('__')
|
2017-01-18 00:43:28 -04:00
|
|
|
|
2016-11-20 22:01:29 -04:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
class _GenericAlias(_Final, _root=True):
|
|
|
|
"""The central part of internal API.
|
2016-10-29 12:54:56 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
This represents a generic version of type 'origin' with type arguments 'params'.
|
|
|
|
There are two kind of these aliases: user defined and special. The special ones
|
|
|
|
are wrappers around builtin collections and ABCs in collections.abc. These must
|
|
|
|
have 'name' always set. If 'inst' is False, then the alias can't be instantiated,
|
|
|
|
this is used by e.g. typing.List and typing.Dict.
|
2016-10-29 12:54:56 -03:00
|
|
|
"""
|
2018-01-20 07:23:59 -04:00
|
|
|
def __init__(self, origin, params, *, inst=True, special=False, name=None):
|
|
|
|
self._inst = inst
|
|
|
|
self._special = special
|
|
|
|
if special and name is None:
|
|
|
|
orig_name = origin.__name__
|
2018-04-04 21:25:15 -03:00
|
|
|
name = _normalize_alias.get(orig_name, orig_name)
|
2018-01-20 07:23:59 -04:00
|
|
|
self._name = name
|
|
|
|
if not isinstance(params, tuple):
|
|
|
|
params = (params,)
|
2016-10-29 12:54:56 -03:00
|
|
|
self.__origin__ = origin
|
|
|
|
self.__args__ = tuple(... if a is _TypingEllipsis else
|
|
|
|
() if a is _TypingEmpty else
|
2018-01-20 07:23:59 -04:00
|
|
|
a for a in params)
|
|
|
|
self.__parameters__ = _collect_type_vars(params)
|
|
|
|
self.__slots__ = None # This is not documented.
|
|
|
|
if not name:
|
|
|
|
self.__module__ = origin.__module__
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2016-09-27 19:20:12 -03:00
|
|
|
@_tp_cache
|
2015-05-22 14:14:11 -03:00
|
|
|
def __getitem__(self, params):
|
2018-01-20 07:23:59 -04:00
|
|
|
if self.__origin__ in (Generic, _Protocol):
|
|
|
|
# Can't subscript Generic[...] or _Protocol[...].
|
|
|
|
raise TypeError(f"Cannot subscript already-subscripted {self}")
|
2015-05-22 14:14:11 -03:00
|
|
|
if not isinstance(params, tuple):
|
|
|
|
params = (params,)
|
|
|
|
msg = "Parameters to generic types must be types."
|
|
|
|
params = tuple(_type_check(p, msg) for p in params)
|
2018-01-20 07:23:59 -04:00
|
|
|
_check_generic(self, params)
|
|
|
|
return _subs_tvars(self, self.__parameters__, params)
|
2017-02-23 23:03:28 -04:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def copy_with(self, params):
|
|
|
|
# We don't copy self._special.
|
|
|
|
return _GenericAlias(self.__origin__, params, name=self._name, inst=self._inst)
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def __repr__(self):
|
|
|
|
if (self._name != 'Callable' or
|
|
|
|
len(self.__args__) == 2 and self.__args__[0] is Ellipsis):
|
|
|
|
if self._name:
|
|
|
|
name = 'typing.' + self._name
|
|
|
|
else:
|
|
|
|
name = _type_repr(self.__origin__)
|
|
|
|
if not self._special:
|
|
|
|
args = f'[{", ".join([_type_repr(a) for a in self.__args__])}]'
|
|
|
|
else:
|
|
|
|
args = ''
|
|
|
|
return (f'{name}{args}')
|
|
|
|
if self._special:
|
|
|
|
return 'typing.Callable'
|
|
|
|
return (f'typing.Callable'
|
|
|
|
f'[[{", ".join([_type_repr(a) for a in self.__args__[:-1]])}], '
|
|
|
|
f'{_type_repr(self.__args__[-1])}]')
|
2016-04-05 12:28:52 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def __eq__(self, other):
|
|
|
|
if not isinstance(other, _GenericAlias):
|
|
|
|
return NotImplemented
|
|
|
|
if self.__origin__ != other.__origin__:
|
|
|
|
return False
|
|
|
|
if self.__origin__ is Union and other.__origin__ is Union:
|
|
|
|
return frozenset(self.__args__) == frozenset(other.__args__)
|
|
|
|
return self.__args__ == other.__args__
|
2016-04-05 12:28:52 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def __hash__(self):
|
|
|
|
if self.__origin__ is Union:
|
|
|
|
return hash((Union, frozenset(self.__args__)))
|
|
|
|
return hash((self.__origin__, self.__args__))
|
|
|
|
|
|
|
|
def __call__(self, *args, **kwargs):
|
|
|
|
if not self._inst:
|
|
|
|
raise TypeError(f"Type {self._name} cannot be instantiated; "
|
|
|
|
f"use {self._name.lower()}() instead")
|
|
|
|
result = self.__origin__(*args, **kwargs)
|
2016-10-29 12:54:56 -03:00
|
|
|
try:
|
2018-01-20 07:23:59 -04:00
|
|
|
result.__orig_class__ = self
|
2016-10-29 12:54:56 -03:00
|
|
|
except AttributeError:
|
|
|
|
pass
|
2018-01-20 07:23:59 -04:00
|
|
|
return result
|
|
|
|
|
|
|
|
def __mro_entries__(self, bases):
|
|
|
|
if self._name: # generic version of an ABC or built-in class
|
|
|
|
res = []
|
|
|
|
if self.__origin__ not in bases:
|
|
|
|
res.append(self.__origin__)
|
|
|
|
i = bases.index(self)
|
|
|
|
if not any(isinstance(b, _GenericAlias) or issubclass(b, Generic)
|
|
|
|
for b in bases[i+1:]):
|
|
|
|
res.append(Generic)
|
|
|
|
return tuple(res)
|
|
|
|
if self.__origin__ is Generic:
|
|
|
|
i = bases.index(self)
|
|
|
|
for b in bases[i+1:]:
|
|
|
|
if isinstance(b, _GenericAlias) and b is not self:
|
|
|
|
return ()
|
|
|
|
return (self.__origin__,)
|
|
|
|
|
|
|
|
def __getattr__(self, attr):
|
2018-04-20 17:08:45 -03:00
|
|
|
# We are careful for copy and pickle.
|
2018-01-20 07:23:59 -04:00
|
|
|
# Also for simplicity we just don't relay all dunder names
|
|
|
|
if '__origin__' in self.__dict__ and not _is_dunder(attr):
|
|
|
|
return getattr(self.__origin__, attr)
|
|
|
|
raise AttributeError(attr)
|
|
|
|
|
|
|
|
def __setattr__(self, attr, val):
|
|
|
|
if _is_dunder(attr) or attr in ('_name', '_inst', '_special'):
|
|
|
|
super().__setattr__(attr, val)
|
|
|
|
else:
|
|
|
|
setattr(self.__origin__, attr, val)
|
2016-10-29 12:54:56 -03:00
|
|
|
|
|
|
|
def __instancecheck__(self, obj):
|
2018-01-20 07:23:59 -04:00
|
|
|
return self.__subclasscheck__(type(obj))
|
2016-10-29 12:54:56 -03:00
|
|
|
|
|
|
|
def __subclasscheck__(self, cls):
|
2018-01-20 07:23:59 -04:00
|
|
|
if self._special:
|
|
|
|
if not isinstance(cls, _GenericAlias):
|
|
|
|
return issubclass(cls, self.__origin__)
|
|
|
|
if cls._special:
|
|
|
|
return issubclass(cls.__origin__, self.__origin__)
|
|
|
|
raise TypeError("Subscripted generics cannot be used with"
|
|
|
|
" class and instance checks")
|
|
|
|
|
2018-03-26 19:01:12 -03:00
|
|
|
def __reduce__(self):
|
|
|
|
if self._special:
|
|
|
|
return self._name
|
2018-05-26 15:19:26 -03:00
|
|
|
|
|
|
|
if self._name:
|
|
|
|
origin = globals()[self._name]
|
|
|
|
else:
|
|
|
|
origin = self.__origin__
|
|
|
|
if (origin is Callable and
|
|
|
|
not (len(self.__args__) == 2 and self.__args__[0] is Ellipsis)):
|
|
|
|
args = list(self.__args__[:-1]), self.__args__[-1]
|
|
|
|
else:
|
|
|
|
args = tuple(self.__args__)
|
|
|
|
if len(args) == 1 and not isinstance(args[0], tuple):
|
|
|
|
args, = args
|
|
|
|
return operator.getitem, (origin, args)
|
2018-03-26 19:01:12 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
|
|
|
|
class _VariadicGenericAlias(_GenericAlias, _root=True):
|
|
|
|
"""Same as _GenericAlias above but for variadic aliases. Currently,
|
|
|
|
this is used only by special internal aliases: Tuple and Callable.
|
2016-10-29 12:54:56 -03:00
|
|
|
"""
|
2018-01-20 07:23:59 -04:00
|
|
|
def __getitem__(self, params):
|
|
|
|
if self._name != 'Callable' or not self._special:
|
|
|
|
return self.__getitem_inner__(params)
|
|
|
|
if not isinstance(params, tuple) or len(params) != 2:
|
2016-10-29 12:54:56 -03:00
|
|
|
raise TypeError("Callable must be used as "
|
|
|
|
"Callable[[arg, ...], result].")
|
2018-01-20 07:23:59 -04:00
|
|
|
args, result = params
|
2016-11-09 17:12:51 -04:00
|
|
|
if args is Ellipsis:
|
2018-01-20 07:23:59 -04:00
|
|
|
params = (Ellipsis, result)
|
2016-04-05 12:28:52 -03:00
|
|
|
else:
|
2016-10-29 12:54:56 -03:00
|
|
|
if not isinstance(args, list):
|
2018-01-20 07:23:59 -04:00
|
|
|
raise TypeError(f"Callable[args, result]: args must be a list."
|
|
|
|
f" Got {args}")
|
|
|
|
params = (tuple(args), result)
|
|
|
|
return self.__getitem_inner__(params)
|
2016-10-29 12:54:56 -03:00
|
|
|
|
|
|
|
@_tp_cache
|
2018-01-20 07:23:59 -04:00
|
|
|
def __getitem_inner__(self, params):
|
|
|
|
if self.__origin__ is tuple and self._special:
|
|
|
|
if params == ():
|
|
|
|
return self.copy_with((_TypingEmpty,))
|
|
|
|
if not isinstance(params, tuple):
|
|
|
|
params = (params,)
|
|
|
|
if len(params) == 2 and params[1] is ...:
|
|
|
|
msg = "Tuple[t, ...]: t must be a type."
|
|
|
|
p = _type_check(params[0], msg)
|
|
|
|
return self.copy_with((p, _TypingEllipsis))
|
|
|
|
msg = "Tuple[t0, t1, ...]: each t must be a type."
|
|
|
|
params = tuple(_type_check(p, msg) for p in params)
|
|
|
|
return self.copy_with(params)
|
|
|
|
if self.__origin__ is collections.abc.Callable and self._special:
|
|
|
|
args, result = params
|
|
|
|
msg = "Callable[args, result]: result must be a type."
|
|
|
|
result = _type_check(result, msg)
|
|
|
|
if args is Ellipsis:
|
|
|
|
return self.copy_with((_TypingEllipsis, result))
|
|
|
|
msg = "Callable[[arg, ...], result]: each arg must be a type."
|
|
|
|
args = tuple(_type_check(arg, msg) for arg in args)
|
|
|
|
params = args + (result,)
|
|
|
|
return self.copy_with(params)
|
|
|
|
return super().__getitem__(params)
|
|
|
|
|
|
|
|
|
|
|
|
class Generic:
|
|
|
|
"""Abstract base class for generic types.
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
A generic type is typically declared by inheriting from
|
|
|
|
this class parameterized with one or more type variables.
|
|
|
|
For example, a generic mapping type might be defined as::
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
class Mapping(Generic[KT, VT]):
|
|
|
|
def __getitem__(self, key: KT) -> VT:
|
|
|
|
...
|
|
|
|
# Etc.
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
This class can then be used as follows::
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def lookup_name(mapping: Mapping[KT, VT], key: KT, default: VT) -> VT:
|
|
|
|
try:
|
|
|
|
return mapping[key]
|
|
|
|
except KeyError:
|
|
|
|
return default
|
2016-09-11 19:34:56 -03:00
|
|
|
"""
|
2018-01-20 07:23:59 -04:00
|
|
|
__slots__ = ()
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def __new__(cls, *args, **kwds):
|
|
|
|
if cls is Generic:
|
|
|
|
raise TypeError("Type Generic cannot be instantiated; "
|
|
|
|
"it can be used only as a base class")
|
2018-05-11 00:10:10 -03:00
|
|
|
if super().__new__ is object.__new__ and cls.__init__ is not object.__init__:
|
2018-05-08 22:23:46 -03:00
|
|
|
obj = super().__new__(cls)
|
|
|
|
else:
|
|
|
|
obj = super().__new__(cls, *args, **kwds)
|
|
|
|
return obj
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
@_tp_cache
|
|
|
|
def __class_getitem__(cls, params):
|
|
|
|
if not isinstance(params, tuple):
|
|
|
|
params = (params,)
|
|
|
|
if not params and cls is not Tuple:
|
|
|
|
raise TypeError(
|
|
|
|
f"Parameter list to {cls.__qualname__}[...] cannot be empty")
|
|
|
|
msg = "Parameters to generic types must be types."
|
|
|
|
params = tuple(_type_check(p, msg) for p in params)
|
|
|
|
if cls is Generic:
|
|
|
|
# Generic can only be subscripted with unique type variables.
|
|
|
|
if not all(isinstance(p, TypeVar) for p in params):
|
|
|
|
raise TypeError(
|
|
|
|
"Parameters to Generic[...] must all be type variables")
|
|
|
|
if len(set(params)) != len(params):
|
|
|
|
raise TypeError(
|
|
|
|
"Parameters to Generic[...] must all be unique")
|
|
|
|
elif cls is _Protocol:
|
|
|
|
# _Protocol is internal at the moment, just skip the check
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
# Subscripting a regular Generic subclass.
|
|
|
|
_check_generic(cls, params)
|
|
|
|
return _GenericAlias(cls, params)
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def __init_subclass__(cls, *args, **kwargs):
|
2018-04-04 13:00:15 -03:00
|
|
|
super().__init_subclass__(*args, **kwargs)
|
2018-01-20 07:23:59 -04:00
|
|
|
tvars = []
|
|
|
|
if '__orig_bases__' in cls.__dict__:
|
|
|
|
error = Generic in cls.__orig_bases__
|
|
|
|
else:
|
|
|
|
error = Generic in cls.__bases__ and cls.__name__ != '_Protocol'
|
|
|
|
if error:
|
|
|
|
raise TypeError("Cannot inherit from plain Generic")
|
|
|
|
if '__orig_bases__' in cls.__dict__:
|
|
|
|
tvars = _collect_type_vars(cls.__orig_bases__)
|
|
|
|
# Look for Generic[T1, ..., Tn].
|
|
|
|
# If found, tvars must be a subset of it.
|
|
|
|
# If not found, tvars is it.
|
|
|
|
# Also check for and reject plain Generic,
|
|
|
|
# and reject multiple Generic[...].
|
|
|
|
gvars = None
|
|
|
|
for base in cls.__orig_bases__:
|
|
|
|
if (isinstance(base, _GenericAlias) and
|
|
|
|
base.__origin__ is Generic):
|
|
|
|
if gvars is not None:
|
|
|
|
raise TypeError(
|
|
|
|
"Cannot inherit from Generic[...] multiple types.")
|
|
|
|
gvars = base.__parameters__
|
|
|
|
if gvars is None:
|
|
|
|
gvars = tvars
|
|
|
|
else:
|
|
|
|
tvarset = set(tvars)
|
|
|
|
gvarset = set(gvars)
|
|
|
|
if not tvarset <= gvarset:
|
|
|
|
s_vars = ', '.join(str(t) for t in tvars if t not in gvarset)
|
|
|
|
s_args = ', '.join(str(g) for g in gvars)
|
|
|
|
raise TypeError(f"Some type variables ({s_vars}) are"
|
|
|
|
f" not listed in Generic[{s_args}]")
|
|
|
|
tvars = gvars
|
|
|
|
cls.__parameters__ = tuple(tvars)
|
2016-09-11 19:34:56 -03:00
|
|
|
|
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
class _TypingEmpty:
|
|
|
|
"""Internal placeholder for () or []. Used by TupleMeta and CallableMeta
|
|
|
|
to allow empty list/tuple in specific places, without allowing them
|
|
|
|
to sneak in where prohibited.
|
|
|
|
"""
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2016-10-29 12:54:56 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
class _TypingEllipsis:
|
|
|
|
"""Internal placeholder for ... (ellipsis)."""
|
2016-09-11 19:34:56 -03:00
|
|
|
|
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
def cast(typ, val):
|
|
|
|
"""Cast a value to a type.
|
|
|
|
|
|
|
|
This returns the value unchanged. To the type checker this
|
|
|
|
signals that the return value has the designated type, but at
|
|
|
|
runtime we intentionally don't check anything (we want this
|
|
|
|
to be as fast as possible).
|
|
|
|
"""
|
|
|
|
return val
|
|
|
|
|
|
|
|
|
|
|
|
def _get_defaults(func):
|
|
|
|
"""Internal helper to extract the default arguments, by name."""
|
2016-11-09 17:12:51 -04:00
|
|
|
try:
|
|
|
|
code = func.__code__
|
|
|
|
except AttributeError:
|
|
|
|
# Some built-in functions don't have __code__, __defaults__, etc.
|
|
|
|
return {}
|
2015-05-22 14:14:11 -03:00
|
|
|
pos_count = code.co_argcount
|
|
|
|
arg_names = code.co_varnames
|
|
|
|
arg_names = arg_names[:pos_count]
|
|
|
|
defaults = func.__defaults__ or ()
|
|
|
|
kwdefaults = func.__kwdefaults__
|
|
|
|
res = dict(kwdefaults) if kwdefaults else {}
|
|
|
|
pos_offset = pos_count - len(defaults)
|
|
|
|
for name, value in zip(arg_names[pos_offset:], defaults):
|
|
|
|
assert name not in res
|
|
|
|
res[name] = value
|
|
|
|
return res
|
|
|
|
|
|
|
|
|
2017-02-13 17:50:14 -04:00
|
|
|
_allowed_types = (types.FunctionType, types.BuiltinFunctionType,
|
|
|
|
types.MethodType, types.ModuleType,
|
2017-05-02 14:14:07 -03:00
|
|
|
WrapperDescriptorType, MethodWrapperType, MethodDescriptorType)
|
2017-02-13 17:50:14 -04:00
|
|
|
|
|
|
|
|
2016-11-09 17:12:51 -04:00
|
|
|
def get_type_hints(obj, globalns=None, localns=None):
|
|
|
|
"""Return type hints for an object.
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2016-11-09 17:12:51 -04:00
|
|
|
This is often the same as obj.__annotations__, but it handles
|
|
|
|
forward references encoded as string literals, and if necessary
|
|
|
|
adds Optional[t] if a default value equal to None is set.
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2016-11-09 17:12:51 -04:00
|
|
|
The argument may be a module, class, method, or function. The annotations
|
|
|
|
are returned as a dictionary. For classes, annotations include also
|
|
|
|
inherited members.
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2016-11-09 17:12:51 -04:00
|
|
|
TypeError is raised if the argument is not of a type that can contain
|
|
|
|
annotations, and an empty dictionary is returned if no annotations are
|
|
|
|
present.
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2016-11-09 17:12:51 -04:00
|
|
|
BEWARE -- the behavior of globalns and localns is counterintuitive
|
|
|
|
(unless you are familiar with how eval() and exec() work). The
|
|
|
|
search order is locals first, then globals.
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2016-11-09 17:12:51 -04:00
|
|
|
- If no dict arguments are passed, an attempt is made to use the
|
2017-09-14 15:33:00 -03:00
|
|
|
globals from obj (or the respective module's globals for classes),
|
|
|
|
and these are also used as the locals. If the object does not appear
|
|
|
|
to have globals, an empty dictionary is used.
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2016-11-09 17:12:51 -04:00
|
|
|
- If one dict argument is passed, it is used for both globals and
|
|
|
|
locals.
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2016-11-09 17:12:51 -04:00
|
|
|
- If two dict arguments are passed, they specify globals and
|
|
|
|
locals, respectively.
|
|
|
|
"""
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2016-11-09 17:12:51 -04:00
|
|
|
if getattr(obj, '__no_type_check__', None):
|
|
|
|
return {}
|
|
|
|
# Classes require a special treatment.
|
|
|
|
if isinstance(obj, type):
|
|
|
|
hints = {}
|
|
|
|
for base in reversed(obj.__mro__):
|
2017-09-14 15:33:00 -03:00
|
|
|
if globalns is None:
|
|
|
|
base_globals = sys.modules[base.__module__].__dict__
|
|
|
|
else:
|
|
|
|
base_globals = globalns
|
2016-11-09 17:12:51 -04:00
|
|
|
ann = base.__dict__.get('__annotations__', {})
|
|
|
|
for name, value in ann.items():
|
2016-09-11 19:34:56 -03:00
|
|
|
if value is None:
|
|
|
|
value = type(None)
|
|
|
|
if isinstance(value, str):
|
2018-05-23 00:32:10 -03:00
|
|
|
value = ForwardRef(value, is_argument=False)
|
2017-09-14 15:33:00 -03:00
|
|
|
value = _eval_type(value, base_globals, localns)
|
2016-09-11 19:34:56 -03:00
|
|
|
hints[name] = value
|
|
|
|
return hints
|
2017-09-14 15:33:00 -03:00
|
|
|
|
|
|
|
if globalns is None:
|
|
|
|
if isinstance(obj, types.ModuleType):
|
|
|
|
globalns = obj.__dict__
|
|
|
|
else:
|
|
|
|
globalns = getattr(obj, '__globals__', {})
|
|
|
|
if localns is None:
|
|
|
|
localns = globalns
|
|
|
|
elif localns is None:
|
|
|
|
localns = globalns
|
2016-11-09 17:12:51 -04:00
|
|
|
hints = getattr(obj, '__annotations__', None)
|
|
|
|
if hints is None:
|
|
|
|
# Return empty annotations for something that _could_ have them.
|
2017-02-13 17:50:14 -04:00
|
|
|
if isinstance(obj, _allowed_types):
|
2016-11-09 17:12:51 -04:00
|
|
|
return {}
|
|
|
|
else:
|
|
|
|
raise TypeError('{!r} is not a module, class, method, '
|
|
|
|
'or function.'.format(obj))
|
|
|
|
defaults = _get_defaults(obj)
|
|
|
|
hints = dict(hints)
|
|
|
|
for name, value in hints.items():
|
|
|
|
if value is None:
|
|
|
|
value = type(None)
|
|
|
|
if isinstance(value, str):
|
2018-01-20 07:23:59 -04:00
|
|
|
value = ForwardRef(value)
|
2016-11-09 17:12:51 -04:00
|
|
|
value = _eval_type(value, globalns, localns)
|
|
|
|
if name in defaults and defaults[name] is None:
|
|
|
|
value = Optional[value]
|
|
|
|
hints[name] = value
|
|
|
|
return hints
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
|
|
|
|
def no_type_check(arg):
|
|
|
|
"""Decorator to indicate that annotations are not type hints.
|
|
|
|
|
|
|
|
The argument must be a class or function; if it is a class, it
|
2016-09-11 19:34:56 -03:00
|
|
|
applies recursively to all methods and classes defined in that class
|
|
|
|
(but not to methods defined in its superclasses or subclasses).
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2016-09-11 19:34:56 -03:00
|
|
|
This mutates the function(s) or class(es) in place.
|
2015-05-22 14:14:11 -03:00
|
|
|
"""
|
|
|
|
if isinstance(arg, type):
|
2016-09-11 19:34:56 -03:00
|
|
|
arg_attrs = arg.__dict__.copy()
|
|
|
|
for attr, val in arg.__dict__.items():
|
2017-09-13 20:25:15 -03:00
|
|
|
if val in arg.__bases__ + (arg,):
|
2016-09-11 19:34:56 -03:00
|
|
|
arg_attrs.pop(attr)
|
|
|
|
for obj in arg_attrs.values():
|
2015-05-22 14:14:11 -03:00
|
|
|
if isinstance(obj, types.FunctionType):
|
|
|
|
obj.__no_type_check__ = True
|
2016-09-11 19:34:56 -03:00
|
|
|
if isinstance(obj, type):
|
|
|
|
no_type_check(obj)
|
|
|
|
try:
|
2015-05-22 14:14:11 -03:00
|
|
|
arg.__no_type_check__ = True
|
2017-01-22 21:43:53 -04:00
|
|
|
except TypeError: # built-in classes
|
2016-09-11 19:34:56 -03:00
|
|
|
pass
|
2015-05-22 14:14:11 -03:00
|
|
|
return arg
|
|
|
|
|
|
|
|
|
|
|
|
def no_type_check_decorator(decorator):
|
|
|
|
"""Decorator to give another decorator the @no_type_check effect.
|
|
|
|
|
|
|
|
This wraps the decorator with something that wraps the decorated
|
|
|
|
function in @no_type_check.
|
|
|
|
"""
|
|
|
|
|
|
|
|
@functools.wraps(decorator)
|
|
|
|
def wrapped_decorator(*args, **kwds):
|
|
|
|
func = decorator(*args, **kwds)
|
|
|
|
func = no_type_check(func)
|
|
|
|
return func
|
|
|
|
|
|
|
|
return wrapped_decorator
|
|
|
|
|
|
|
|
|
2016-04-05 12:28:52 -03:00
|
|
|
def _overload_dummy(*args, **kwds):
|
|
|
|
"""Helper for @overload to raise when called."""
|
|
|
|
raise NotImplementedError(
|
|
|
|
"You should not call an overloaded function. "
|
|
|
|
"A series of @overload-decorated functions "
|
|
|
|
"outside a stub module should always be followed "
|
|
|
|
"by an implementation that is not @overload-ed.")
|
|
|
|
|
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
def overload(func):
|
2016-04-05 12:28:52 -03:00
|
|
|
"""Decorator for overloaded functions/methods.
|
|
|
|
|
|
|
|
In a stub file, place two or more stub definitions for the same
|
|
|
|
function in a row, each decorated with @overload. For example:
|
|
|
|
|
|
|
|
@overload
|
|
|
|
def utf8(value: None) -> None: ...
|
|
|
|
@overload
|
|
|
|
def utf8(value: bytes) -> bytes: ...
|
|
|
|
@overload
|
|
|
|
def utf8(value: str) -> bytes: ...
|
|
|
|
|
|
|
|
In a non-stub file (i.e. a regular .py file), do the same but
|
|
|
|
follow it with an implementation. The implementation should *not*
|
|
|
|
be decorated with @overload. For example:
|
|
|
|
|
|
|
|
@overload
|
|
|
|
def utf8(value: None) -> None: ...
|
|
|
|
@overload
|
|
|
|
def utf8(value: bytes) -> bytes: ...
|
|
|
|
@overload
|
|
|
|
def utf8(value: str) -> bytes: ...
|
|
|
|
def utf8(value):
|
|
|
|
# implementation goes here
|
|
|
|
"""
|
|
|
|
return _overload_dummy
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
class _ProtocolMeta(type):
|
2015-05-22 14:14:11 -03:00
|
|
|
"""Internal metaclass for _Protocol.
|
|
|
|
|
|
|
|
This exists so _Protocol classes can be generic without deriving
|
|
|
|
from Generic.
|
|
|
|
"""
|
|
|
|
|
2015-08-05 07:11:06 -03:00
|
|
|
def __instancecheck__(self, obj):
|
2016-11-19 14:32:41 -04:00
|
|
|
if _Protocol not in self.__bases__:
|
|
|
|
return super().__instancecheck__(obj)
|
2015-08-05 07:11:06 -03:00
|
|
|
raise TypeError("Protocols cannot be used with isinstance().")
|
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
def __subclasscheck__(self, cls):
|
|
|
|
if not self._is_protocol:
|
|
|
|
# No structural checks since this isn't a protocol.
|
|
|
|
return NotImplemented
|
|
|
|
|
|
|
|
if self is _Protocol:
|
|
|
|
# Every class is a subclass of the empty protocol.
|
|
|
|
return True
|
|
|
|
|
|
|
|
# Find all attributes defined in the protocol.
|
|
|
|
attrs = self._get_protocol_attrs()
|
|
|
|
|
|
|
|
for attr in attrs:
|
|
|
|
if not any(attr in d.__dict__ for d in cls.__mro__):
|
|
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
|
|
def _get_protocol_attrs(self):
|
|
|
|
# Get all Protocol base classes.
|
|
|
|
protocol_bases = []
|
|
|
|
for c in self.__mro__:
|
|
|
|
if getattr(c, '_is_protocol', False) and c.__name__ != '_Protocol':
|
|
|
|
protocol_bases.append(c)
|
|
|
|
|
|
|
|
# Get attributes included in protocol.
|
|
|
|
attrs = set()
|
|
|
|
for base in protocol_bases:
|
|
|
|
for attr in base.__dict__.keys():
|
|
|
|
# Include attributes not defined in any non-protocol bases.
|
|
|
|
for c in self.__mro__:
|
|
|
|
if (c is not base and attr in c.__dict__ and
|
|
|
|
not getattr(c, '_is_protocol', False)):
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
if (not attr.startswith('_abc_') and
|
2016-04-05 12:28:52 -03:00
|
|
|
attr != '__abstractmethods__' and
|
2016-09-11 19:34:56 -03:00
|
|
|
attr != '__annotations__' and
|
|
|
|
attr != '__weakref__' and
|
2016-04-05 12:28:52 -03:00
|
|
|
attr != '_is_protocol' and
|
2017-09-13 20:25:15 -03:00
|
|
|
attr != '_gorg' and
|
2016-04-05 12:28:52 -03:00
|
|
|
attr != '__dict__' and
|
|
|
|
attr != '__args__' and
|
|
|
|
attr != '__slots__' and
|
|
|
|
attr != '_get_protocol_attrs' and
|
|
|
|
attr != '__next_in_mro__' and
|
|
|
|
attr != '__parameters__' and
|
|
|
|
attr != '__origin__' and
|
2016-10-21 18:27:58 -03:00
|
|
|
attr != '__orig_bases__' and
|
2016-05-18 12:35:00 -03:00
|
|
|
attr != '__extra__' and
|
2016-10-29 12:54:56 -03:00
|
|
|
attr != '__tree_hash__' and
|
2016-04-05 12:28:52 -03:00
|
|
|
attr != '__module__'):
|
2015-05-22 14:14:11 -03:00
|
|
|
attrs.add(attr)
|
|
|
|
|
|
|
|
return attrs
|
|
|
|
|
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
class _Protocol(Generic, metaclass=_ProtocolMeta):
|
2015-05-22 14:14:11 -03:00
|
|
|
"""Internal base class for protocol classes.
|
|
|
|
|
2016-11-20 22:01:29 -04:00
|
|
|
This implements a simple-minded structural issubclass check
|
2015-05-22 14:14:11 -03:00
|
|
|
(similar but more general than the one-offs in collections.abc
|
|
|
|
such as Hashable).
|
|
|
|
"""
|
|
|
|
|
2015-08-05 07:11:06 -03:00
|
|
|
__slots__ = ()
|
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
_is_protocol = True
|
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
def __class_getitem__(cls, params):
|
|
|
|
return super().__class_getitem__(params)
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
# Some unconstrained type variables. These are used by the container types.
|
|
|
|
# (These are not for export.)
|
|
|
|
T = TypeVar('T') # Any type.
|
|
|
|
KT = TypeVar('KT') # Key type.
|
|
|
|
VT = TypeVar('VT') # Value type.
|
|
|
|
T_co = TypeVar('T_co', covariant=True) # Any type covariant containers.
|
|
|
|
V_co = TypeVar('V_co', covariant=True) # Any type covariant containers.
|
|
|
|
VT_co = TypeVar('VT_co', covariant=True) # Value type covariant containers.
|
|
|
|
T_contra = TypeVar('T_contra', contravariant=True) # Ditto contravariant.
|
|
|
|
# Internal type variable used for Type[].
|
|
|
|
CT_co = TypeVar('CT_co', covariant=True, bound=type)
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
# A useful type variable with constraints. This represents string types.
|
|
|
|
# (This one *is* for export!)
|
|
|
|
AnyStr = TypeVar('AnyStr', bytes, str)
|
2016-10-29 20:05:26 -03:00
|
|
|
|
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
# Various ABCs mimicking those in collections.abc.
|
|
|
|
def _alias(origin, params, inst=True):
|
|
|
|
return _GenericAlias(origin, params, special=True, inst=inst)
|
|
|
|
|
|
|
|
Hashable = _alias(collections.abc.Hashable, ()) # Not generic.
|
|
|
|
Awaitable = _alias(collections.abc.Awaitable, T_co)
|
|
|
|
Coroutine = _alias(collections.abc.Coroutine, (T_co, T_contra, V_co))
|
|
|
|
AsyncIterable = _alias(collections.abc.AsyncIterable, T_co)
|
|
|
|
AsyncIterator = _alias(collections.abc.AsyncIterator, T_co)
|
|
|
|
Iterable = _alias(collections.abc.Iterable, T_co)
|
|
|
|
Iterator = _alias(collections.abc.Iterator, T_co)
|
|
|
|
Reversible = _alias(collections.abc.Reversible, T_co)
|
|
|
|
Sized = _alias(collections.abc.Sized, ()) # Not generic.
|
|
|
|
Container = _alias(collections.abc.Container, T_co)
|
|
|
|
Collection = _alias(collections.abc.Collection, T_co)
|
|
|
|
Callable = _VariadicGenericAlias(collections.abc.Callable, (), special=True)
|
|
|
|
Callable.__doc__ = \
|
|
|
|
"""Callable type; Callable[[int], str] is a function of (int) -> str.
|
2016-10-29 20:05:26 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
The subscription syntax must always be used with exactly two
|
|
|
|
values: the argument list and the return type. The argument list
|
|
|
|
must be a list of types or ellipsis; the return type must be a single type.
|
2015-12-03 21:31:24 -04:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
There is no syntax to indicate optional or keyword arguments,
|
|
|
|
such function types are rarely used as callback types.
|
|
|
|
"""
|
|
|
|
AbstractSet = _alias(collections.abc.Set, T_co)
|
|
|
|
MutableSet = _alias(collections.abc.MutableSet, T)
|
|
|
|
# NOTE: Mapping is only covariant in the value type.
|
|
|
|
Mapping = _alias(collections.abc.Mapping, (KT, VT_co))
|
|
|
|
MutableMapping = _alias(collections.abc.MutableMapping, (KT, VT))
|
|
|
|
Sequence = _alias(collections.abc.Sequence, T_co)
|
|
|
|
MutableSequence = _alias(collections.abc.MutableSequence, T)
|
|
|
|
ByteString = _alias(collections.abc.ByteString, ()) # Not generic
|
|
|
|
Tuple = _VariadicGenericAlias(tuple, (), inst=False, special=True)
|
|
|
|
Tuple.__doc__ = \
|
|
|
|
"""Tuple type; Tuple[X, Y] is the cross-product type of X and Y.
|
2015-12-03 21:31:24 -04:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
Example: Tuple[T1, T2] is a tuple of two elements corresponding
|
|
|
|
to type variables T1 and T2. Tuple[int, float, str] is a tuple
|
|
|
|
of an int, a float and a string.
|
2015-12-03 21:31:24 -04:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
To specify a variable-length tuple of homogeneous type, use Tuple[T, ...].
|
|
|
|
"""
|
|
|
|
List = _alias(list, T, inst=False)
|
|
|
|
Deque = _alias(collections.deque, T)
|
|
|
|
Set = _alias(set, T, inst=False)
|
|
|
|
FrozenSet = _alias(frozenset, T_co, inst=False)
|
|
|
|
MappingView = _alias(collections.abc.MappingView, T_co)
|
|
|
|
KeysView = _alias(collections.abc.KeysView, KT)
|
|
|
|
ItemsView = _alias(collections.abc.ItemsView, (KT, VT_co))
|
|
|
|
ValuesView = _alias(collections.abc.ValuesView, VT_co)
|
|
|
|
ContextManager = _alias(contextlib.AbstractContextManager, T_co)
|
|
|
|
AsyncContextManager = _alias(contextlib.AbstractAsyncContextManager, T_co)
|
|
|
|
Dict = _alias(dict, (KT, VT), inst=False)
|
|
|
|
DefaultDict = _alias(collections.defaultdict, (KT, VT))
|
2018-12-02 11:53:14 -04:00
|
|
|
OrderedDict = _alias(collections.OrderedDict, (KT, VT))
|
2018-01-20 07:23:59 -04:00
|
|
|
Counter = _alias(collections.Counter, T)
|
|
|
|
ChainMap = _alias(collections.ChainMap, (KT, VT))
|
|
|
|
Generator = _alias(collections.abc.Generator, (T_co, T_contra, V_co))
|
|
|
|
AsyncGenerator = _alias(collections.abc.AsyncGenerator, (T_co, T_contra))
|
|
|
|
Type = _alias(type, CT_co, inst=False)
|
|
|
|
Type.__doc__ = \
|
|
|
|
"""A special construct usable to annotate class objects.
|
2015-12-03 21:31:24 -04:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
For example, suppose we have the following classes::
|
2016-04-05 12:28:52 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
class User: ... # Abstract base for User classes
|
|
|
|
class BasicUser(User): ...
|
|
|
|
class ProUser(User): ...
|
|
|
|
class TeamUser(User): ...
|
2015-12-03 21:31:24 -04:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
And a function that takes a class argument that's a subclass of
|
|
|
|
User and returns an instance of the corresponding class::
|
2015-12-03 21:31:24 -04:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
U = TypeVar('U', bound=User)
|
|
|
|
def new_user(user_class: Type[U]) -> U:
|
|
|
|
user = user_class()
|
|
|
|
# (Here we could write the user object to a database)
|
|
|
|
return user
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
joe = new_user(BasicUser)
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
At this point the type checker knows that joe has type BasicUser.
|
|
|
|
"""
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
|
|
|
|
class SupportsInt(_Protocol):
|
2015-08-05 07:11:06 -03:00
|
|
|
__slots__ = ()
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def __int__(self) -> int:
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class SupportsFloat(_Protocol):
|
2015-08-05 07:11:06 -03:00
|
|
|
__slots__ = ()
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def __float__(self) -> float:
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class SupportsComplex(_Protocol):
|
2015-08-05 07:11:06 -03:00
|
|
|
__slots__ = ()
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def __complex__(self) -> complex:
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class SupportsBytes(_Protocol):
|
2015-08-05 07:11:06 -03:00
|
|
|
__slots__ = ()
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def __bytes__(self) -> bytes:
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
2015-08-05 07:11:06 -03:00
|
|
|
class SupportsAbs(_Protocol[T_co]):
|
|
|
|
__slots__ = ()
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
@abstractmethod
|
2015-08-05 07:11:06 -03:00
|
|
|
def __abs__(self) -> T_co:
|
2015-05-22 14:14:11 -03:00
|
|
|
pass
|
|
|
|
|
|
|
|
|
2015-08-05 07:11:06 -03:00
|
|
|
class SupportsRound(_Protocol[T_co]):
|
|
|
|
__slots__ = ()
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
@abstractmethod
|
2015-08-05 07:11:06 -03:00
|
|
|
def __round__(self, ndigits: int = 0) -> T_co:
|
2015-05-22 14:14:11 -03:00
|
|
|
pass
|
|
|
|
|
|
|
|
|
2016-09-11 19:34:56 -03:00
|
|
|
def _make_nmtuple(name, types):
|
2016-11-15 13:48:06 -04:00
|
|
|
msg = "NamedTuple('Name', [(f0, t0), (f1, t1), ...]); each t must be a type"
|
|
|
|
types = [(n, _type_check(t, msg)) for n, t in types]
|
2016-09-11 19:34:56 -03:00
|
|
|
nm_tpl = collections.namedtuple(name, [n for n, t in types])
|
2017-01-18 00:43:28 -04:00
|
|
|
# Prior to PEP 526, only _field_types attribute was assigned.
|
2019-03-18 13:53:56 -03:00
|
|
|
# Now __annotations__ are used and _field_types is deprecated (remove in 3.9)
|
|
|
|
nm_tpl.__annotations__ = nm_tpl._field_types = dict(types)
|
2016-09-11 19:34:56 -03:00
|
|
|
try:
|
|
|
|
nm_tpl.__module__ = sys._getframe(2).f_globals.get('__name__', '__main__')
|
|
|
|
except (AttributeError, ValueError):
|
|
|
|
pass
|
|
|
|
return nm_tpl
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
|
2017-02-13 17:50:14 -04:00
|
|
|
# attributes prohibited to set in NamedTuple class syntax
|
|
|
|
_prohibited = ('__new__', '__init__', '__slots__', '__getnewargs__',
|
|
|
|
'_fields', '_field_defaults', '_field_types',
|
2017-05-02 14:14:07 -03:00
|
|
|
'_make', '_replace', '_asdict', '_source')
|
2017-02-13 17:50:14 -04:00
|
|
|
|
|
|
|
_special = ('__module__', '__name__', '__qualname__', '__annotations__')
|
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2016-11-15 13:48:06 -04:00
|
|
|
class NamedTupleMeta(type):
|
2015-05-22 14:14:11 -03:00
|
|
|
|
2016-11-15 13:48:06 -04:00
|
|
|
def __new__(cls, typename, bases, ns):
|
|
|
|
if ns.get('_root', False):
|
|
|
|
return super().__new__(cls, typename, bases, ns)
|
|
|
|
types = ns.get('__annotations__', {})
|
2017-01-18 12:03:50 -04:00
|
|
|
nm_tpl = _make_nmtuple(typename, types.items())
|
|
|
|
defaults = []
|
|
|
|
defaults_dict = {}
|
|
|
|
for field_name in types:
|
|
|
|
if field_name in ns:
|
|
|
|
default_value = ns[field_name]
|
|
|
|
defaults.append(default_value)
|
|
|
|
defaults_dict[field_name] = default_value
|
|
|
|
elif defaults:
|
2017-01-22 21:43:53 -04:00
|
|
|
raise TypeError("Non-default namedtuple field {field_name} cannot "
|
|
|
|
"follow default field(s) {default_names}"
|
2017-01-18 12:03:50 -04:00
|
|
|
.format(field_name=field_name,
|
|
|
|
default_names=', '.join(defaults_dict.keys())))
|
2019-03-18 13:53:56 -03:00
|
|
|
nm_tpl.__new__.__annotations__ = dict(types)
|
2017-01-18 12:03:50 -04:00
|
|
|
nm_tpl.__new__.__defaults__ = tuple(defaults)
|
|
|
|
nm_tpl._field_defaults = defaults_dict
|
2017-01-22 21:47:20 -04:00
|
|
|
# update from user namespace without overriding special namedtuple attributes
|
|
|
|
for key in ns:
|
2017-02-13 17:50:14 -04:00
|
|
|
if key in _prohibited:
|
|
|
|
raise AttributeError("Cannot overwrite NamedTuple attribute " + key)
|
|
|
|
elif key not in _special and key not in nm_tpl._fields:
|
2017-01-22 21:47:20 -04:00
|
|
|
setattr(nm_tpl, key, ns[key])
|
2017-01-18 12:03:50 -04:00
|
|
|
return nm_tpl
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2017-01-22 21:43:53 -04:00
|
|
|
|
2016-11-15 13:48:06 -04:00
|
|
|
class NamedTuple(metaclass=NamedTupleMeta):
|
|
|
|
"""Typed version of namedtuple.
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2016-11-15 13:48:06 -04:00
|
|
|
Usage in Python versions >= 3.6::
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2016-11-15 13:48:06 -04:00
|
|
|
class Employee(NamedTuple):
|
|
|
|
name: str
|
|
|
|
id: int
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2016-11-15 13:48:06 -04:00
|
|
|
This is equivalent to::
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2016-11-15 13:48:06 -04:00
|
|
|
Employee = collections.namedtuple('Employee', ['name', 'id'])
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2019-03-18 13:53:56 -03:00
|
|
|
The resulting class has an extra __annotations__ attribute, giving a
|
|
|
|
dict that maps field names to types. (The field names are also in
|
|
|
|
the _fields attribute, which is part of the namedtuple API.)
|
|
|
|
Alternative equivalent keyword syntax is also accepted::
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2016-11-15 13:48:06 -04:00
|
|
|
Employee = NamedTuple('Employee', name=str, id=int)
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2016-11-15 13:48:06 -04:00
|
|
|
In Python versions <= 3.5 use::
|
2016-09-11 19:34:56 -03:00
|
|
|
|
2016-11-15 13:48:06 -04:00
|
|
|
Employee = NamedTuple('Employee', [('name', str), ('id', int)])
|
|
|
|
"""
|
|
|
|
_root = True
|
|
|
|
|
|
|
|
def __new__(self, typename, fields=None, **kwargs):
|
|
|
|
if fields is None:
|
|
|
|
fields = kwargs.items()
|
|
|
|
elif kwargs:
|
|
|
|
raise TypeError("Either list of fields or keywords"
|
|
|
|
" can be provided to NamedTuple, not both")
|
2016-09-11 19:34:56 -03:00
|
|
|
return _make_nmtuple(typename, fields)
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
|
2016-06-08 15:19:11 -03:00
|
|
|
def NewType(name, tp):
|
|
|
|
"""NewType creates simple unique types with almost zero
|
|
|
|
runtime overhead. NewType(name, tp) is considered a subtype of tp
|
|
|
|
by static type checkers. At runtime, NewType(name, tp) returns
|
|
|
|
a dummy function that simply returns its argument. Usage::
|
|
|
|
|
|
|
|
UserId = NewType('UserId', int)
|
|
|
|
|
|
|
|
def name_by_id(user_id: UserId) -> str:
|
|
|
|
...
|
|
|
|
|
|
|
|
UserId('user') # Fails type check
|
|
|
|
|
|
|
|
name_by_id(42) # Fails type check
|
|
|
|
name_by_id(UserId(42)) # OK
|
|
|
|
|
|
|
|
num = UserId(5) + 1 # type: int
|
|
|
|
"""
|
|
|
|
|
|
|
|
def new_type(x):
|
|
|
|
return x
|
|
|
|
|
|
|
|
new_type.__name__ = name
|
|
|
|
new_type.__supertype__ = tp
|
|
|
|
return new_type
|
|
|
|
|
|
|
|
|
2016-04-05 18:54:25 -03:00
|
|
|
# Python-version-specific alias (Python 2: unicode; Python 3: str)
|
|
|
|
Text = str
|
|
|
|
|
|
|
|
|
2016-06-08 15:19:11 -03:00
|
|
|
# Constant that's True when type checking, but False here.
|
|
|
|
TYPE_CHECKING = False
|
|
|
|
|
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
class IO(Generic[AnyStr]):
|
|
|
|
"""Generic base class for TextIO and BinaryIO.
|
|
|
|
|
|
|
|
This is an abstract, generic version of the return of open().
|
|
|
|
|
|
|
|
NOTE: This does not distinguish between the different possible
|
|
|
|
classes (text vs. binary, read vs. write vs. read/write,
|
|
|
|
append-only, unbuffered). The TextIO and BinaryIO subclasses
|
|
|
|
below capture the distinctions between text vs. binary, which is
|
|
|
|
pervasive in the interface; however we currently do not offer a
|
|
|
|
way to track the other distinctions in the type system.
|
|
|
|
"""
|
|
|
|
|
2015-08-05 07:11:06 -03:00
|
|
|
__slots__ = ()
|
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
@abstractproperty
|
|
|
|
def mode(self) -> str:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractproperty
|
|
|
|
def name(self) -> str:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def close(self) -> None:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def closed(self) -> bool:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def fileno(self) -> int:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def flush(self) -> None:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def isatty(self) -> bool:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def read(self, n: int = -1) -> AnyStr:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def readable(self) -> bool:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def readline(self, limit: int = -1) -> AnyStr:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def readlines(self, hint: int = -1) -> List[AnyStr]:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def seek(self, offset: int, whence: int = 0) -> int:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def seekable(self) -> bool:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def tell(self) -> int:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def truncate(self, size: int = None) -> int:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def writable(self) -> bool:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def write(self, s: AnyStr) -> int:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def writelines(self, lines: List[AnyStr]) -> None:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def __enter__(self) -> 'IO[AnyStr]':
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def __exit__(self, type, value, traceback) -> None:
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class BinaryIO(IO[bytes]):
|
|
|
|
"""Typed version of the return of open() in binary mode."""
|
|
|
|
|
2015-08-05 07:11:06 -03:00
|
|
|
__slots__ = ()
|
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
@abstractmethod
|
|
|
|
def write(self, s: Union[bytes, bytearray]) -> int:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def __enter__(self) -> 'BinaryIO':
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class TextIO(IO[str]):
|
|
|
|
"""Typed version of the return of open() in text mode."""
|
|
|
|
|
2015-08-05 07:11:06 -03:00
|
|
|
__slots__ = ()
|
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
@abstractproperty
|
|
|
|
def buffer(self) -> BinaryIO:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractproperty
|
|
|
|
def encoding(self) -> str:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractproperty
|
2016-11-09 17:12:51 -04:00
|
|
|
def errors(self) -> Optional[str]:
|
2015-05-22 14:14:11 -03:00
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractproperty
|
|
|
|
def line_buffering(self) -> bool:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractproperty
|
|
|
|
def newlines(self) -> Any:
|
|
|
|
pass
|
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def __enter__(self) -> 'TextIO':
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class io:
|
|
|
|
"""Wrapper namespace for IO generic classes."""
|
|
|
|
|
|
|
|
__all__ = ['IO', 'TextIO', 'BinaryIO']
|
|
|
|
IO = IO
|
|
|
|
TextIO = TextIO
|
|
|
|
BinaryIO = BinaryIO
|
|
|
|
|
2017-01-22 21:43:53 -04:00
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
io.__name__ = __name__ + '.io'
|
|
|
|
sys.modules[io.__name__] = io
|
|
|
|
|
2018-01-20 07:23:59 -04:00
|
|
|
Pattern = _alias(stdlib_re.Pattern, AnyStr)
|
|
|
|
Match = _alias(stdlib_re.Match, AnyStr)
|
2015-05-22 14:14:11 -03:00
|
|
|
|
|
|
|
class re:
|
|
|
|
"""Wrapper namespace for re type aliases."""
|
|
|
|
|
|
|
|
__all__ = ['Pattern', 'Match']
|
|
|
|
Pattern = Pattern
|
|
|
|
Match = Match
|
|
|
|
|
2017-01-22 21:43:53 -04:00
|
|
|
|
2015-05-22 14:14:11 -03:00
|
|
|
re.__name__ = __name__ + '.re'
|
|
|
|
sys.modules[re.__name__] = re
|