Backed out changeset 57776eee74f2
This commit is contained in:
parent
1c858c352b
commit
8252cc9832
203
Lib/functools.py
203
Lib/functools.py
|
@ -419,129 +419,120 @@ def lru_cache(maxsize=128, typed=False):
|
||||||
if maxsize is not None and not isinstance(maxsize, int):
|
if maxsize is not None and not isinstance(maxsize, int):
|
||||||
raise TypeError('Expected maxsize to be an integer or None')
|
raise TypeError('Expected maxsize to be an integer or None')
|
||||||
|
|
||||||
def decorating_function(user_function):
|
|
||||||
wrapper = _lru_cache_wrapper(user_function, maxsize, typed, _CacheInfo)
|
|
||||||
return update_wrapper(wrapper, user_function)
|
|
||||||
|
|
||||||
return decorating_function
|
|
||||||
|
|
||||||
def _lru_cache_wrapper(user_function, maxsize, typed, _CacheInfo):
|
|
||||||
# Constants shared by all lru cache instances:
|
# Constants shared by all lru cache instances:
|
||||||
sentinel = object() # unique object used to signal cache misses
|
sentinel = object() # unique object used to signal cache misses
|
||||||
make_key = _make_key # build a key from the function arguments
|
make_key = _make_key # build a key from the function arguments
|
||||||
PREV, NEXT, KEY, RESULT = 0, 1, 2, 3 # names for the link fields
|
PREV, NEXT, KEY, RESULT = 0, 1, 2, 3 # names for the link fields
|
||||||
|
|
||||||
cache = {}
|
def decorating_function(user_function):
|
||||||
hits = misses = 0
|
cache = {}
|
||||||
full = False
|
hits = misses = 0
|
||||||
cache_get = cache.get # bound method to lookup a key or return None
|
full = False
|
||||||
lock = RLock() # because linkedlist updates aren't threadsafe
|
cache_get = cache.get # bound method to lookup a key or return None
|
||||||
root = [] # root of the circular doubly linked list
|
lock = RLock() # because linkedlist updates aren't threadsafe
|
||||||
root[:] = [root, root, None, None] # initialize by pointing to self
|
root = [] # root of the circular doubly linked list
|
||||||
|
root[:] = [root, root, None, None] # initialize by pointing to self
|
||||||
|
|
||||||
if maxsize == 0:
|
if maxsize == 0:
|
||||||
|
|
||||||
def wrapper(*args, **kwds):
|
def wrapper(*args, **kwds):
|
||||||
# No caching -- just a statistics update after a successful call
|
# No caching -- just a statistics update after a successful call
|
||||||
nonlocal misses
|
nonlocal misses
|
||||||
result = user_function(*args, **kwds)
|
result = user_function(*args, **kwds)
|
||||||
misses += 1
|
misses += 1
|
||||||
return result
|
|
||||||
|
|
||||||
elif maxsize is None:
|
|
||||||
|
|
||||||
def wrapper(*args, **kwds):
|
|
||||||
# Simple caching without ordering or size limit
|
|
||||||
nonlocal hits, misses
|
|
||||||
key = make_key(args, kwds, typed)
|
|
||||||
result = cache_get(key, sentinel)
|
|
||||||
if result is not sentinel:
|
|
||||||
hits += 1
|
|
||||||
return result
|
return result
|
||||||
result = user_function(*args, **kwds)
|
|
||||||
cache[key] = result
|
|
||||||
misses += 1
|
|
||||||
return result
|
|
||||||
|
|
||||||
else:
|
elif maxsize is None:
|
||||||
|
|
||||||
def wrapper(*args, **kwds):
|
def wrapper(*args, **kwds):
|
||||||
# Size limited caching that tracks accesses by recency
|
# Simple caching without ordering or size limit
|
||||||
nonlocal root, hits, misses, full
|
nonlocal hits, misses
|
||||||
key = make_key(args, kwds, typed)
|
key = make_key(args, kwds, typed)
|
||||||
with lock:
|
result = cache_get(key, sentinel)
|
||||||
link = cache_get(key)
|
if result is not sentinel:
|
||||||
if link is not None:
|
|
||||||
# Move the link to the front of the circular queue
|
|
||||||
link_prev, link_next, _key, result = link
|
|
||||||
link_prev[NEXT] = link_next
|
|
||||||
link_next[PREV] = link_prev
|
|
||||||
last = root[PREV]
|
|
||||||
last[NEXT] = root[PREV] = link
|
|
||||||
link[PREV] = last
|
|
||||||
link[NEXT] = root
|
|
||||||
hits += 1
|
hits += 1
|
||||||
return result
|
return result
|
||||||
result = user_function(*args, **kwds)
|
result = user_function(*args, **kwds)
|
||||||
with lock:
|
cache[key] = result
|
||||||
if key in cache:
|
|
||||||
# Getting here means that this same key was added to the
|
|
||||||
# cache while the lock was released. Since the link
|
|
||||||
# update is already done, we need only return the
|
|
||||||
# computed result and update the count of misses.
|
|
||||||
pass
|
|
||||||
elif full:
|
|
||||||
# Use the old root to store the new key and result.
|
|
||||||
oldroot = root
|
|
||||||
oldroot[KEY] = key
|
|
||||||
oldroot[RESULT] = result
|
|
||||||
# Empty the oldest link and make it the new root.
|
|
||||||
# Keep a reference to the old key and old result to
|
|
||||||
# prevent their ref counts from going to zero during the
|
|
||||||
# update. That will prevent potentially arbitrary object
|
|
||||||
# clean-up code (i.e. __del__) from running while we're
|
|
||||||
# still adjusting the links.
|
|
||||||
root = oldroot[NEXT]
|
|
||||||
oldkey = root[KEY]
|
|
||||||
oldresult = root[RESULT]
|
|
||||||
root[KEY] = root[RESULT] = None
|
|
||||||
# Now update the cache dictionary.
|
|
||||||
del cache[oldkey]
|
|
||||||
# Save the potentially reentrant cache[key] assignment
|
|
||||||
# for last, after the root and links have been put in
|
|
||||||
# a consistent state.
|
|
||||||
cache[key] = oldroot
|
|
||||||
else:
|
|
||||||
# Put result in a new link at the front of the queue.
|
|
||||||
last = root[PREV]
|
|
||||||
link = [last, root, key, result]
|
|
||||||
last[NEXT] = root[PREV] = cache[key] = link
|
|
||||||
full = (len(cache) >= maxsize)
|
|
||||||
misses += 1
|
misses += 1
|
||||||
return result
|
return result
|
||||||
|
|
||||||
def cache_info():
|
else:
|
||||||
"""Report cache statistics"""
|
|
||||||
with lock:
|
|
||||||
return _CacheInfo(hits, misses, maxsize, len(cache))
|
|
||||||
|
|
||||||
def cache_clear():
|
def wrapper(*args, **kwds):
|
||||||
"""Clear the cache and cache statistics"""
|
# Size limited caching that tracks accesses by recency
|
||||||
nonlocal hits, misses, full
|
nonlocal root, hits, misses, full
|
||||||
with lock:
|
key = make_key(args, kwds, typed)
|
||||||
cache.clear()
|
with lock:
|
||||||
root[:] = [root, root, None, None]
|
link = cache_get(key)
|
||||||
hits = misses = 0
|
if link is not None:
|
||||||
full = False
|
# Move the link to the front of the circular queue
|
||||||
|
link_prev, link_next, _key, result = link
|
||||||
|
link_prev[NEXT] = link_next
|
||||||
|
link_next[PREV] = link_prev
|
||||||
|
last = root[PREV]
|
||||||
|
last[NEXT] = root[PREV] = link
|
||||||
|
link[PREV] = last
|
||||||
|
link[NEXT] = root
|
||||||
|
hits += 1
|
||||||
|
return result
|
||||||
|
result = user_function(*args, **kwds)
|
||||||
|
with lock:
|
||||||
|
if key in cache:
|
||||||
|
# Getting here means that this same key was added to the
|
||||||
|
# cache while the lock was released. Since the link
|
||||||
|
# update is already done, we need only return the
|
||||||
|
# computed result and update the count of misses.
|
||||||
|
pass
|
||||||
|
elif full:
|
||||||
|
# Use the old root to store the new key and result.
|
||||||
|
oldroot = root
|
||||||
|
oldroot[KEY] = key
|
||||||
|
oldroot[RESULT] = result
|
||||||
|
# Empty the oldest link and make it the new root.
|
||||||
|
# Keep a reference to the old key and old result to
|
||||||
|
# prevent their ref counts from going to zero during the
|
||||||
|
# update. That will prevent potentially arbitrary object
|
||||||
|
# clean-up code (i.e. __del__) from running while we're
|
||||||
|
# still adjusting the links.
|
||||||
|
root = oldroot[NEXT]
|
||||||
|
oldkey = root[KEY]
|
||||||
|
oldresult = root[RESULT]
|
||||||
|
root[KEY] = root[RESULT] = None
|
||||||
|
# Now update the cache dictionary.
|
||||||
|
del cache[oldkey]
|
||||||
|
# Save the potentially reentrant cache[key] assignment
|
||||||
|
# for last, after the root and links have been put in
|
||||||
|
# a consistent state.
|
||||||
|
cache[key] = oldroot
|
||||||
|
else:
|
||||||
|
# Put result in a new link at the front of the queue.
|
||||||
|
last = root[PREV]
|
||||||
|
link = [last, root, key, result]
|
||||||
|
last[NEXT] = root[PREV] = cache[key] = link
|
||||||
|
full = (len(cache) >= maxsize)
|
||||||
|
misses += 1
|
||||||
|
return result
|
||||||
|
|
||||||
wrapper.cache_info = cache_info
|
def cache_info():
|
||||||
wrapper.cache_clear = cache_clear
|
"""Report cache statistics"""
|
||||||
return update_wrapper(wrapper, user_function)
|
with lock:
|
||||||
|
return _CacheInfo(hits, misses, maxsize, len(cache))
|
||||||
|
|
||||||
try:
|
def cache_clear():
|
||||||
from _functools import _lru_cache_wrapper
|
"""Clear the cache and cache statistics"""
|
||||||
except ImportError:
|
nonlocal hits, misses, full
|
||||||
pass
|
with lock:
|
||||||
|
cache.clear()
|
||||||
|
root[:] = [root, root, None, None]
|
||||||
|
hits = misses = 0
|
||||||
|
full = False
|
||||||
|
|
||||||
|
wrapper.cache_info = cache_info
|
||||||
|
wrapper.cache_clear = cache_clear
|
||||||
|
return update_wrapper(wrapper, user_function)
|
||||||
|
|
||||||
|
return decorating_function
|
||||||
|
|
||||||
|
|
||||||
################################################################################
|
################################################################################
|
||||||
|
|
|
@ -7,10 +7,6 @@ import sys
|
||||||
from test import support
|
from test import support
|
||||||
import unittest
|
import unittest
|
||||||
from weakref import proxy
|
from weakref import proxy
|
||||||
try:
|
|
||||||
import threading
|
|
||||||
except ImportError:
|
|
||||||
threading = None
|
|
||||||
|
|
||||||
import functools
|
import functools
|
||||||
|
|
||||||
|
@ -916,12 +912,12 @@ class Orderable_LT:
|
||||||
return self.value == other.value
|
return self.value == other.value
|
||||||
|
|
||||||
|
|
||||||
class TestLRU:
|
class TestLRU(unittest.TestCase):
|
||||||
|
|
||||||
def test_lru(self):
|
def test_lru(self):
|
||||||
def orig(x, y):
|
def orig(x, y):
|
||||||
return 3 * x + y
|
return 3 * x + y
|
||||||
f = self.module.lru_cache(maxsize=20)(orig)
|
f = functools.lru_cache(maxsize=20)(orig)
|
||||||
hits, misses, maxsize, currsize = f.cache_info()
|
hits, misses, maxsize, currsize = f.cache_info()
|
||||||
self.assertEqual(maxsize, 20)
|
self.assertEqual(maxsize, 20)
|
||||||
self.assertEqual(currsize, 0)
|
self.assertEqual(currsize, 0)
|
||||||
|
@ -959,7 +955,7 @@ class TestLRU:
|
||||||
self.assertEqual(currsize, 1)
|
self.assertEqual(currsize, 1)
|
||||||
|
|
||||||
# test size zero (which means "never-cache")
|
# test size zero (which means "never-cache")
|
||||||
@self.module.lru_cache(0)
|
@functools.lru_cache(0)
|
||||||
def f():
|
def f():
|
||||||
nonlocal f_cnt
|
nonlocal f_cnt
|
||||||
f_cnt += 1
|
f_cnt += 1
|
||||||
|
@ -975,7 +971,7 @@ class TestLRU:
|
||||||
self.assertEqual(currsize, 0)
|
self.assertEqual(currsize, 0)
|
||||||
|
|
||||||
# test size one
|
# test size one
|
||||||
@self.module.lru_cache(1)
|
@functools.lru_cache(1)
|
||||||
def f():
|
def f():
|
||||||
nonlocal f_cnt
|
nonlocal f_cnt
|
||||||
f_cnt += 1
|
f_cnt += 1
|
||||||
|
@ -991,7 +987,7 @@ class TestLRU:
|
||||||
self.assertEqual(currsize, 1)
|
self.assertEqual(currsize, 1)
|
||||||
|
|
||||||
# test size two
|
# test size two
|
||||||
@self.module.lru_cache(2)
|
@functools.lru_cache(2)
|
||||||
def f(x):
|
def f(x):
|
||||||
nonlocal f_cnt
|
nonlocal f_cnt
|
||||||
f_cnt += 1
|
f_cnt += 1
|
||||||
|
@ -1008,7 +1004,7 @@ class TestLRU:
|
||||||
self.assertEqual(currsize, 2)
|
self.assertEqual(currsize, 2)
|
||||||
|
|
||||||
def test_lru_with_maxsize_none(self):
|
def test_lru_with_maxsize_none(self):
|
||||||
@self.module.lru_cache(maxsize=None)
|
@functools.lru_cache(maxsize=None)
|
||||||
def fib(n):
|
def fib(n):
|
||||||
if n < 2:
|
if n < 2:
|
||||||
return n
|
return n
|
||||||
|
@ -1016,26 +1012,17 @@ class TestLRU:
|
||||||
self.assertEqual([fib(n) for n in range(16)],
|
self.assertEqual([fib(n) for n in range(16)],
|
||||||
[0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, 89, 144, 233, 377, 610])
|
[0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, 89, 144, 233, 377, 610])
|
||||||
self.assertEqual(fib.cache_info(),
|
self.assertEqual(fib.cache_info(),
|
||||||
self.module._CacheInfo(hits=28, misses=16, maxsize=None, currsize=16))
|
functools._CacheInfo(hits=28, misses=16, maxsize=None, currsize=16))
|
||||||
fib.cache_clear()
|
fib.cache_clear()
|
||||||
self.assertEqual(fib.cache_info(),
|
self.assertEqual(fib.cache_info(),
|
||||||
self.module._CacheInfo(hits=0, misses=0, maxsize=None, currsize=0))
|
functools._CacheInfo(hits=0, misses=0, maxsize=None, currsize=0))
|
||||||
|
|
||||||
def test_lru_with_maxsize_negative(self):
|
|
||||||
@self.module.lru_cache(maxsize=-10)
|
|
||||||
def eq(n):
|
|
||||||
return n
|
|
||||||
for i in (0, 1):
|
|
||||||
self.assertEqual([eq(n) for n in range(150)], list(range(150)))
|
|
||||||
self.assertEqual(eq.cache_info(),
|
|
||||||
self.module._CacheInfo(hits=0, misses=300, maxsize=-10, currsize=1))
|
|
||||||
|
|
||||||
def test_lru_with_exceptions(self):
|
def test_lru_with_exceptions(self):
|
||||||
# Verify that user_function exceptions get passed through without
|
# Verify that user_function exceptions get passed through without
|
||||||
# creating a hard-to-read chained exception.
|
# creating a hard-to-read chained exception.
|
||||||
# http://bugs.python.org/issue13177
|
# http://bugs.python.org/issue13177
|
||||||
for maxsize in (None, 128):
|
for maxsize in (None, 128):
|
||||||
@self.module.lru_cache(maxsize)
|
@functools.lru_cache(maxsize)
|
||||||
def func(i):
|
def func(i):
|
||||||
return 'abc'[i]
|
return 'abc'[i]
|
||||||
self.assertEqual(func(0), 'a')
|
self.assertEqual(func(0), 'a')
|
||||||
|
@ -1048,7 +1035,7 @@ class TestLRU:
|
||||||
|
|
||||||
def test_lru_with_types(self):
|
def test_lru_with_types(self):
|
||||||
for maxsize in (None, 128):
|
for maxsize in (None, 128):
|
||||||
@self.module.lru_cache(maxsize=maxsize, typed=True)
|
@functools.lru_cache(maxsize=maxsize, typed=True)
|
||||||
def square(x):
|
def square(x):
|
||||||
return x * x
|
return x * x
|
||||||
self.assertEqual(square(3), 9)
|
self.assertEqual(square(3), 9)
|
||||||
|
@ -1063,7 +1050,7 @@ class TestLRU:
|
||||||
self.assertEqual(square.cache_info().misses, 4)
|
self.assertEqual(square.cache_info().misses, 4)
|
||||||
|
|
||||||
def test_lru_with_keyword_args(self):
|
def test_lru_with_keyword_args(self):
|
||||||
@self.module.lru_cache()
|
@functools.lru_cache()
|
||||||
def fib(n):
|
def fib(n):
|
||||||
if n < 2:
|
if n < 2:
|
||||||
return n
|
return n
|
||||||
|
@ -1073,13 +1060,13 @@ class TestLRU:
|
||||||
[0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, 89, 144, 233, 377, 610]
|
[0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, 89, 144, 233, 377, 610]
|
||||||
)
|
)
|
||||||
self.assertEqual(fib.cache_info(),
|
self.assertEqual(fib.cache_info(),
|
||||||
self.module._CacheInfo(hits=28, misses=16, maxsize=128, currsize=16))
|
functools._CacheInfo(hits=28, misses=16, maxsize=128, currsize=16))
|
||||||
fib.cache_clear()
|
fib.cache_clear()
|
||||||
self.assertEqual(fib.cache_info(),
|
self.assertEqual(fib.cache_info(),
|
||||||
self.module._CacheInfo(hits=0, misses=0, maxsize=128, currsize=0))
|
functools._CacheInfo(hits=0, misses=0, maxsize=128, currsize=0))
|
||||||
|
|
||||||
def test_lru_with_keyword_args_maxsize_none(self):
|
def test_lru_with_keyword_args_maxsize_none(self):
|
||||||
@self.module.lru_cache(maxsize=None)
|
@functools.lru_cache(maxsize=None)
|
||||||
def fib(n):
|
def fib(n):
|
||||||
if n < 2:
|
if n < 2:
|
||||||
return n
|
return n
|
||||||
|
@ -1087,71 +1074,15 @@ class TestLRU:
|
||||||
self.assertEqual([fib(n=number) for number in range(16)],
|
self.assertEqual([fib(n=number) for number in range(16)],
|
||||||
[0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, 89, 144, 233, 377, 610])
|
[0, 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, 89, 144, 233, 377, 610])
|
||||||
self.assertEqual(fib.cache_info(),
|
self.assertEqual(fib.cache_info(),
|
||||||
self.module._CacheInfo(hits=28, misses=16, maxsize=None, currsize=16))
|
functools._CacheInfo(hits=28, misses=16, maxsize=None, currsize=16))
|
||||||
fib.cache_clear()
|
fib.cache_clear()
|
||||||
self.assertEqual(fib.cache_info(),
|
self.assertEqual(fib.cache_info(),
|
||||||
self.module._CacheInfo(hits=0, misses=0, maxsize=None, currsize=0))
|
functools._CacheInfo(hits=0, misses=0, maxsize=None, currsize=0))
|
||||||
|
|
||||||
def test_lru_cache_decoration(self):
|
|
||||||
def f(zomg: 'zomg_annotation'):
|
|
||||||
"""f doc string"""
|
|
||||||
return 42
|
|
||||||
g = self.module.lru_cache()(f)
|
|
||||||
for attr in self.module.WRAPPER_ASSIGNMENTS:
|
|
||||||
self.assertEqual(getattr(g, attr), getattr(f, attr))
|
|
||||||
|
|
||||||
@unittest.skipUnless(threading, 'This test requires threading.')
|
|
||||||
def test_lru_cache_threaded(self):
|
|
||||||
def orig(x, y):
|
|
||||||
return 3 * x + y
|
|
||||||
f = self.module.lru_cache(maxsize=20)(orig)
|
|
||||||
hits, misses, maxsize, currsize = f.cache_info()
|
|
||||||
self.assertEqual(currsize, 0)
|
|
||||||
|
|
||||||
def full(f, *args):
|
|
||||||
for _ in range(10):
|
|
||||||
f(*args)
|
|
||||||
|
|
||||||
def clear(f):
|
|
||||||
for _ in range(10):
|
|
||||||
f.cache_clear()
|
|
||||||
|
|
||||||
orig_si = sys.getswitchinterval()
|
|
||||||
sys.setswitchinterval(1e-6)
|
|
||||||
try:
|
|
||||||
# create 5 threads in order to fill cache
|
|
||||||
threads = []
|
|
||||||
for k in range(5):
|
|
||||||
t = threading.Thread(target=full, args=[f, k, k])
|
|
||||||
t.start()
|
|
||||||
threads.append(t)
|
|
||||||
|
|
||||||
for t in threads:
|
|
||||||
t.join()
|
|
||||||
|
|
||||||
hits, misses, maxsize, currsize = f.cache_info()
|
|
||||||
self.assertEqual(hits, 45)
|
|
||||||
self.assertEqual(misses, 5)
|
|
||||||
self.assertEqual(currsize, 5)
|
|
||||||
|
|
||||||
# create 5 threads in order to fill cache and 1 to clear it
|
|
||||||
cleaner = threading.Thread(target=clear, args=[f])
|
|
||||||
cleaner.start()
|
|
||||||
threads = [cleaner]
|
|
||||||
for k in range(5):
|
|
||||||
t = threading.Thread(target=full, args=[f, k, k])
|
|
||||||
t.start()
|
|
||||||
threads.append(t)
|
|
||||||
|
|
||||||
for t in threads:
|
|
||||||
t.join()
|
|
||||||
finally:
|
|
||||||
sys.setswitchinterval(orig_si)
|
|
||||||
|
|
||||||
def test_need_for_rlock(self):
|
def test_need_for_rlock(self):
|
||||||
# This will deadlock on an LRU cache that uses a regular lock
|
# This will deadlock on an LRU cache that uses a regular lock
|
||||||
|
|
||||||
@self.module.lru_cache(maxsize=10)
|
@functools.lru_cache(maxsize=10)
|
||||||
def test_func(x):
|
def test_func(x):
|
||||||
'Used to demonstrate a reentrant lru_cache call within a single thread'
|
'Used to demonstrate a reentrant lru_cache call within a single thread'
|
||||||
return x
|
return x
|
||||||
|
@ -1179,12 +1110,6 @@ class TestLRU:
|
||||||
def f():
|
def f():
|
||||||
pass
|
pass
|
||||||
|
|
||||||
class TestLRUC(TestLRU, unittest.TestCase):
|
|
||||||
module = c_functools
|
|
||||||
|
|
||||||
class TestLRUPy(TestLRU, unittest.TestCase):
|
|
||||||
module = py_functools
|
|
||||||
|
|
||||||
|
|
||||||
class TestSingleDispatch(unittest.TestCase):
|
class TestSingleDispatch(unittest.TestCase):
|
||||||
def test_simple_overloads(self):
|
def test_simple_overloads(self):
|
||||||
|
|
|
@ -63,9 +63,6 @@ Core and Builtins
|
||||||
Library
|
Library
|
||||||
-------
|
-------
|
||||||
|
|
||||||
- Issue #14373: Added C implementation of functools.lru_cache(). Based on
|
|
||||||
patches by Matt Joiner and Alexey Kachayev.
|
|
||||||
|
|
||||||
- Issue 24230: The tempfile module now accepts bytes for prefix, suffix and dir
|
- Issue 24230: The tempfile module now accepts bytes for prefix, suffix and dir
|
||||||
parameters and returns bytes in such situations (matching the os module APIs).
|
parameters and returns bytes in such situations (matching the os module APIs).
|
||||||
|
|
||||||
|
|
|
@ -590,539 +590,6 @@ For example, reduce(lambda x, y: x+y, [1, 2, 3, 4, 5]) calculates\n\
|
||||||
of the sequence in the calculation, and serves as a default when the\n\
|
of the sequence in the calculation, and serves as a default when the\n\
|
||||||
sequence is empty.");
|
sequence is empty.");
|
||||||
|
|
||||||
/* lru_cache object **********************************************************/
|
|
||||||
|
|
||||||
/* this object is used delimit args and keywords in the cache keys */
|
|
||||||
static PyObject *kwd_mark = NULL;
|
|
||||||
|
|
||||||
struct lru_list_elem;
|
|
||||||
struct lru_cache_object;
|
|
||||||
|
|
||||||
typedef struct lru_list_elem {
|
|
||||||
PyObject_HEAD
|
|
||||||
struct lru_list_elem *prev, *next; /* borrowed links */
|
|
||||||
PyObject *key, *result;
|
|
||||||
} lru_list_elem;
|
|
||||||
|
|
||||||
static void
|
|
||||||
lru_list_elem_dealloc(lru_list_elem *link)
|
|
||||||
{
|
|
||||||
_PyObject_GC_UNTRACK(link);
|
|
||||||
Py_XDECREF(link->key);
|
|
||||||
Py_XDECREF(link->result);
|
|
||||||
PyObject_GC_Del(link);
|
|
||||||
}
|
|
||||||
|
|
||||||
static int
|
|
||||||
lru_list_elem_traverse(lru_list_elem *link, visitproc visit, void *arg)
|
|
||||||
{
|
|
||||||
Py_VISIT(link->key);
|
|
||||||
Py_VISIT(link->result);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int
|
|
||||||
lru_list_elem_clear(lru_list_elem *link)
|
|
||||||
{
|
|
||||||
Py_CLEAR(link->key);
|
|
||||||
Py_CLEAR(link->result);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static PyTypeObject lru_list_elem_type = {
|
|
||||||
PyVarObject_HEAD_INIT(&PyType_Type, 0)
|
|
||||||
"functools._lru_list_elem", /* tp_name */
|
|
||||||
sizeof(lru_list_elem), /* tp_basicsize */
|
|
||||||
0, /* tp_itemsize */
|
|
||||||
/* methods */
|
|
||||||
(destructor)lru_list_elem_dealloc, /* tp_dealloc */
|
|
||||||
0, /* tp_print */
|
|
||||||
0, /* tp_getattr */
|
|
||||||
0, /* tp_setattr */
|
|
||||||
0, /* tp_reserved */
|
|
||||||
0, /* tp_repr */
|
|
||||||
0, /* tp_as_number */
|
|
||||||
0, /* tp_as_sequence */
|
|
||||||
0, /* tp_as_mapping */
|
|
||||||
0, /* tp_hash */
|
|
||||||
0, /* tp_call */
|
|
||||||
0, /* tp_str */
|
|
||||||
0, /* tp_getattro */
|
|
||||||
0, /* tp_setattro */
|
|
||||||
0, /* tp_as_buffer */
|
|
||||||
Py_TPFLAGS_DEFAULT | Py_TPFLAGS_HAVE_GC, /* tp_flags */
|
|
||||||
0, /* tp_doc */
|
|
||||||
(traverseproc)lru_list_elem_traverse, /* tp_traverse */
|
|
||||||
(inquiry)lru_list_elem_clear, /* tp_clear */
|
|
||||||
};
|
|
||||||
|
|
||||||
|
|
||||||
typedef PyObject *(*lru_cache_ternaryfunc)(struct lru_cache_object *, PyObject *, PyObject *);
|
|
||||||
|
|
||||||
typedef struct lru_cache_object {
|
|
||||||
lru_list_elem root; /* includes PyObject_HEAD */
|
|
||||||
Py_ssize_t maxsize;
|
|
||||||
PyObject *maxsize_O;
|
|
||||||
PyObject *func;
|
|
||||||
lru_cache_ternaryfunc wrapper;
|
|
||||||
PyObject *cache;
|
|
||||||
PyObject *cache_info_type;
|
|
||||||
Py_ssize_t misses, hits;
|
|
||||||
int typed;
|
|
||||||
PyObject *dict;
|
|
||||||
int full;
|
|
||||||
} lru_cache_object;
|
|
||||||
|
|
||||||
static PyTypeObject lru_cache_type;
|
|
||||||
|
|
||||||
static PyObject *
|
|
||||||
lru_cache_make_key(PyObject *args, PyObject *kwds, int typed)
|
|
||||||
{
|
|
||||||
PyObject *key, *sorted_items;
|
|
||||||
Py_ssize_t key_size, pos, key_pos;
|
|
||||||
|
|
||||||
/* short path, key will match args anyway, which is a tuple */
|
|
||||||
if (!typed && !kwds) {
|
|
||||||
Py_INCREF(args);
|
|
||||||
return args;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (kwds && PyDict_Size(kwds) > 0) {
|
|
||||||
sorted_items = PyDict_Items(kwds);
|
|
||||||
if (!sorted_items)
|
|
||||||
return NULL;
|
|
||||||
if (PyList_Sort(sorted_items) < 0) {
|
|
||||||
Py_DECREF(sorted_items);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
} else
|
|
||||||
sorted_items = NULL;
|
|
||||||
|
|
||||||
key_size = PyTuple_GET_SIZE(args);
|
|
||||||
if (sorted_items)
|
|
||||||
key_size += PyList_GET_SIZE(sorted_items);
|
|
||||||
if (typed)
|
|
||||||
key_size *= 2;
|
|
||||||
if (sorted_items)
|
|
||||||
key_size++;
|
|
||||||
|
|
||||||
key = PyTuple_New(key_size);
|
|
||||||
if (key == NULL)
|
|
||||||
goto done;
|
|
||||||
|
|
||||||
key_pos = 0;
|
|
||||||
for (pos = 0; pos < PyTuple_GET_SIZE(args); ++pos) {
|
|
||||||
PyObject *item = PyTuple_GET_ITEM(args, pos);
|
|
||||||
Py_INCREF(item);
|
|
||||||
PyTuple_SET_ITEM(key, key_pos++, item);
|
|
||||||
}
|
|
||||||
if (sorted_items) {
|
|
||||||
Py_INCREF(kwd_mark);
|
|
||||||
PyTuple_SET_ITEM(key, key_pos++, kwd_mark);
|
|
||||||
for (pos = 0; pos < PyList_GET_SIZE(sorted_items); ++pos) {
|
|
||||||
PyObject *item = PyList_GET_ITEM(sorted_items, pos);
|
|
||||||
Py_INCREF(item);
|
|
||||||
PyTuple_SET_ITEM(key, key_pos++, item);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if (typed) {
|
|
||||||
for (pos = 0; pos < PyTuple_GET_SIZE(args); ++pos) {
|
|
||||||
PyObject *item = (PyObject *)Py_TYPE(PyTuple_GET_ITEM(args, pos));
|
|
||||||
Py_INCREF(item);
|
|
||||||
PyTuple_SET_ITEM(key, key_pos++, item);
|
|
||||||
}
|
|
||||||
if (sorted_items) {
|
|
||||||
for (pos = 0; pos < PyList_GET_SIZE(sorted_items); ++pos) {
|
|
||||||
PyObject *tp_items = PyList_GET_ITEM(sorted_items, pos);
|
|
||||||
PyObject *item = (PyObject *)Py_TYPE(PyTuple_GET_ITEM(tp_items, 1));
|
|
||||||
Py_INCREF(item);
|
|
||||||
PyTuple_SET_ITEM(key, key_pos++, item);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
assert(key_pos == key_size);
|
|
||||||
|
|
||||||
done:
|
|
||||||
if (sorted_items)
|
|
||||||
Py_DECREF(sorted_items);
|
|
||||||
return key;
|
|
||||||
}
|
|
||||||
|
|
||||||
static PyObject *
|
|
||||||
uncached_lru_cache_wrapper(lru_cache_object *self, PyObject *args, PyObject *kwds)
|
|
||||||
{
|
|
||||||
PyObject *result = PyObject_Call(self->func, args, kwds);
|
|
||||||
if (!result)
|
|
||||||
return NULL;
|
|
||||||
self->misses++;
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
|
|
||||||
static PyObject *
|
|
||||||
infinite_lru_cache_wrapper(lru_cache_object *self, PyObject *args, PyObject *kwds)
|
|
||||||
{
|
|
||||||
PyObject *result;
|
|
||||||
PyObject *key = lru_cache_make_key(args, kwds, self->typed);
|
|
||||||
if (!key)
|
|
||||||
return NULL;
|
|
||||||
result = PyDict_GetItemWithError(self->cache, key);
|
|
||||||
if (result) {
|
|
||||||
Py_INCREF(result);
|
|
||||||
self->hits++;
|
|
||||||
Py_DECREF(key);
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
if (PyErr_Occurred()) {
|
|
||||||
Py_DECREF(key);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
result = PyObject_Call(self->func, args, kwds);
|
|
||||||
if (!result) {
|
|
||||||
Py_DECREF(key);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
if (PyDict_SetItem(self->cache, key, result) < 0) {
|
|
||||||
Py_DECREF(result);
|
|
||||||
Py_DECREF(key);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
Py_DECREF(key);
|
|
||||||
self->misses++;
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void
|
|
||||||
lru_cache_extricate_link(lru_list_elem *link)
|
|
||||||
{
|
|
||||||
link->prev->next = link->next;
|
|
||||||
link->next->prev = link->prev;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void
|
|
||||||
lru_cache_append_link(lru_cache_object *self, lru_list_elem *link)
|
|
||||||
{
|
|
||||||
lru_list_elem *root = &self->root;
|
|
||||||
lru_list_elem *last = root->prev;
|
|
||||||
last->next = root->prev = link;
|
|
||||||
link->prev = last;
|
|
||||||
link->next = root;
|
|
||||||
}
|
|
||||||
|
|
||||||
static PyObject *
|
|
||||||
bounded_lru_cache_wrapper(lru_cache_object *self, PyObject *args, PyObject *kwds)
|
|
||||||
{
|
|
||||||
lru_list_elem *link;
|
|
||||||
PyObject *key, *result;
|
|
||||||
|
|
||||||
key = lru_cache_make_key(args, kwds, self->typed);
|
|
||||||
if (!key)
|
|
||||||
return NULL;
|
|
||||||
link = (lru_list_elem *)PyDict_GetItemWithError(self->cache, key);
|
|
||||||
if (link) {
|
|
||||||
lru_cache_extricate_link(link);
|
|
||||||
lru_cache_append_link(self, link);
|
|
||||||
self->hits++;
|
|
||||||
result = link->result;
|
|
||||||
Py_INCREF(result);
|
|
||||||
Py_DECREF(key);
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
if (PyErr_Occurred()) {
|
|
||||||
Py_DECREF(key);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
result = PyObject_Call(self->func, args, kwds);
|
|
||||||
if (!result) {
|
|
||||||
Py_DECREF(key);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
if (self->full && self->root.next != &self->root) {
|
|
||||||
/* Use the oldest item to store the new key and result. */
|
|
||||||
PyObject *oldkey, *oldresult;
|
|
||||||
/* Extricate the oldest item. */
|
|
||||||
link = self->root.next;
|
|
||||||
lru_cache_extricate_link(link);
|
|
||||||
/* Remove it from the cache.
|
|
||||||
The cache dict holds one reference to the link,
|
|
||||||
and the linked list holds yet one reference to it. */
|
|
||||||
if (PyDict_DelItem(self->cache, link->key) < 0) {
|
|
||||||
lru_cache_append_link(self, link);
|
|
||||||
Py_DECREF(key);
|
|
||||||
Py_DECREF(result);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
/* Keep a reference to the old key and old result to
|
|
||||||
prevent their ref counts from going to zero during the
|
|
||||||
update. That will prevent potentially arbitrary object
|
|
||||||
clean-up code (i.e. __del__) from running while we're
|
|
||||||
still adjusting the links. */
|
|
||||||
oldkey = link->key;
|
|
||||||
oldresult = link->result;
|
|
||||||
|
|
||||||
link->key = key;
|
|
||||||
link->result = result;
|
|
||||||
if (PyDict_SetItem(self->cache, key, (PyObject *)link) < 0) {
|
|
||||||
Py_DECREF(link);
|
|
||||||
Py_DECREF(oldkey);
|
|
||||||
Py_DECREF(oldresult);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
lru_cache_append_link(self, link);
|
|
||||||
Py_INCREF(result); /* for return */
|
|
||||||
Py_DECREF(oldkey);
|
|
||||||
Py_DECREF(oldresult);
|
|
||||||
} else {
|
|
||||||
/* Put result in a new link at the front of the queue. */
|
|
||||||
link = (lru_list_elem *)PyObject_GC_New(lru_list_elem,
|
|
||||||
&lru_list_elem_type);
|
|
||||||
if (link == NULL) {
|
|
||||||
Py_DECREF(key);
|
|
||||||
Py_DECREF(result);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
link->key = key;
|
|
||||||
link->result = result;
|
|
||||||
_PyObject_GC_TRACK(link);
|
|
||||||
if (PyDict_SetItem(self->cache, key, (PyObject *)link) < 0) {
|
|
||||||
Py_DECREF(link);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
lru_cache_append_link(self, link);
|
|
||||||
Py_INCREF(result); /* for return */
|
|
||||||
self->full = (PyDict_Size(self->cache) >= self->maxsize);
|
|
||||||
}
|
|
||||||
self->misses++;
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
|
|
||||||
static PyObject *
|
|
||||||
lru_cache_new(PyTypeObject *type, PyObject *args, PyObject *kw)
|
|
||||||
{
|
|
||||||
PyObject *func, *maxsize_O, *cache_info_type;
|
|
||||||
int typed;
|
|
||||||
lru_cache_object *obj;
|
|
||||||
Py_ssize_t maxsize;
|
|
||||||
PyObject *(*wrapper)(lru_cache_object *, PyObject *, PyObject *);
|
|
||||||
static char *keywords[] = {"user_function", "maxsize", "typed",
|
|
||||||
"cache_info_type", NULL};
|
|
||||||
|
|
||||||
if (!PyArg_ParseTupleAndKeywords(args, kw, "OOpO:lru_cache", keywords,
|
|
||||||
&func, &maxsize_O, &typed,
|
|
||||||
&cache_info_type)) {
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!PyCallable_Check(func)) {
|
|
||||||
PyErr_SetString(PyExc_TypeError,
|
|
||||||
"the first argument must be callable");
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* select the caching function, and make/inc maxsize_O */
|
|
||||||
if (maxsize_O == Py_None) {
|
|
||||||
wrapper = infinite_lru_cache_wrapper;
|
|
||||||
/* use this only to initialize lru_cache_object attribute maxsize */
|
|
||||||
maxsize = -1;
|
|
||||||
} else if (PyIndex_Check(maxsize_O)) {
|
|
||||||
maxsize = PyNumber_AsSsize_t(maxsize_O, PyExc_OverflowError);
|
|
||||||
if (maxsize == -1 && PyErr_Occurred())
|
|
||||||
return NULL;
|
|
||||||
if (maxsize == 0)
|
|
||||||
wrapper = uncached_lru_cache_wrapper;
|
|
||||||
else
|
|
||||||
wrapper = bounded_lru_cache_wrapper;
|
|
||||||
} else {
|
|
||||||
PyErr_SetString(PyExc_TypeError, "maxsize should be integer or None");
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
obj = (lru_cache_object *)type->tp_alloc(type, 0);
|
|
||||||
if (obj == NULL)
|
|
||||||
return NULL;
|
|
||||||
|
|
||||||
if (!(obj->cache = PyDict_New())) {
|
|
||||||
Py_DECREF(obj);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
obj->root.prev = &obj->root;
|
|
||||||
obj->root.next = &obj->root;
|
|
||||||
obj->maxsize = maxsize;
|
|
||||||
Py_INCREF(maxsize_O);
|
|
||||||
obj->maxsize_O = maxsize_O;
|
|
||||||
Py_INCREF(func);
|
|
||||||
obj->func = func;
|
|
||||||
obj->wrapper = wrapper;
|
|
||||||
obj->misses = obj->hits = 0;
|
|
||||||
obj->typed = typed;
|
|
||||||
Py_INCREF(cache_info_type);
|
|
||||||
obj->cache_info_type = cache_info_type;
|
|
||||||
|
|
||||||
return (PyObject *)obj;
|
|
||||||
}
|
|
||||||
|
|
||||||
static lru_list_elem *
|
|
||||||
lru_cache_unlink_list(lru_cache_object *self)
|
|
||||||
{
|
|
||||||
lru_list_elem *root = &self->root;
|
|
||||||
lru_list_elem *link = root->next;
|
|
||||||
if (link == root)
|
|
||||||
return NULL;
|
|
||||||
root->prev->next = NULL;
|
|
||||||
root->next = root->prev = root;
|
|
||||||
return link;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void
|
|
||||||
lru_cache_clear_list(lru_list_elem *link)
|
|
||||||
{
|
|
||||||
while (link != NULL) {
|
|
||||||
lru_list_elem *next = link->next;
|
|
||||||
Py_DECREF(link);
|
|
||||||
link = next;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
static void
|
|
||||||
lru_cache_dealloc(lru_cache_object *obj)
|
|
||||||
{
|
|
||||||
lru_list_elem *list = lru_cache_unlink_list(obj);
|
|
||||||
Py_XDECREF(obj->maxsize_O);
|
|
||||||
Py_XDECREF(obj->func);
|
|
||||||
Py_XDECREF(obj->cache);
|
|
||||||
Py_XDECREF(obj->dict);
|
|
||||||
Py_XDECREF(obj->cache_info_type);
|
|
||||||
lru_cache_clear_list(list);
|
|
||||||
Py_TYPE(obj)->tp_free(obj);
|
|
||||||
}
|
|
||||||
|
|
||||||
static PyObject *
|
|
||||||
lru_cache_call(lru_cache_object *self, PyObject *args, PyObject *kwds)
|
|
||||||
{
|
|
||||||
return self->wrapper(self, args, kwds);
|
|
||||||
}
|
|
||||||
|
|
||||||
static PyObject *
|
|
||||||
lru_cache_cache_info(lru_cache_object *self, PyObject *unused)
|
|
||||||
{
|
|
||||||
return PyObject_CallFunction(self->cache_info_type, "nnOn",
|
|
||||||
self->hits, self->misses, self->maxsize_O,
|
|
||||||
PyDict_Size(self->cache));
|
|
||||||
}
|
|
||||||
|
|
||||||
static PyObject *
|
|
||||||
lru_cache_cache_clear(lru_cache_object *self, PyObject *unused)
|
|
||||||
{
|
|
||||||
lru_list_elem *list = lru_cache_unlink_list(self);
|
|
||||||
self->hits = self->misses = 0;
|
|
||||||
self->full = 0;
|
|
||||||
PyDict_Clear(self->cache);
|
|
||||||
lru_cache_clear_list(list);
|
|
||||||
Py_RETURN_NONE;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int
|
|
||||||
lru_cache_tp_traverse(lru_cache_object *self, visitproc visit, void *arg)
|
|
||||||
{
|
|
||||||
lru_list_elem *link = self->root.next;
|
|
||||||
while (link != &self->root) {
|
|
||||||
lru_list_elem *next = link->next;
|
|
||||||
Py_VISIT(link);
|
|
||||||
link = next;
|
|
||||||
}
|
|
||||||
Py_VISIT(self->maxsize_O);
|
|
||||||
Py_VISIT(self->func);
|
|
||||||
Py_VISIT(self->cache);
|
|
||||||
Py_VISIT(self->cache_info_type);
|
|
||||||
Py_VISIT(self->dict);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int
|
|
||||||
lru_cache_tp_clear(lru_cache_object *self)
|
|
||||||
{
|
|
||||||
lru_list_elem *list = lru_cache_unlink_list(self);
|
|
||||||
Py_CLEAR(self->maxsize_O);
|
|
||||||
Py_CLEAR(self->func);
|
|
||||||
Py_CLEAR(self->cache);
|
|
||||||
Py_CLEAR(self->cache_info_type);
|
|
||||||
Py_CLEAR(self->dict);
|
|
||||||
lru_cache_clear_list(list);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
PyDoc_STRVAR(lru_cache_doc,
|
|
||||||
"Create a cached callable that wraps another function.\n\
|
|
||||||
\n\
|
|
||||||
user_function: the function being cached\n\
|
|
||||||
\n\
|
|
||||||
maxsize: 0 for no caching\n\
|
|
||||||
None for unlimited cache size\n\
|
|
||||||
n for a bounded cache\n\
|
|
||||||
\n\
|
|
||||||
typed: False cache f(3) and f(3.0) as identical calls\n\
|
|
||||||
True cache f(3) and f(3.0) as distinct calls\n\
|
|
||||||
\n\
|
|
||||||
cache_info_type: namedtuple class with the fields:\n\
|
|
||||||
hits misses currsize maxsize\n"
|
|
||||||
);
|
|
||||||
|
|
||||||
static PyMethodDef lru_cache_methods[] = {
|
|
||||||
{"cache_info", (PyCFunction)lru_cache_cache_info, METH_NOARGS},
|
|
||||||
{"cache_clear", (PyCFunction)lru_cache_cache_clear, METH_NOARGS},
|
|
||||||
{NULL}
|
|
||||||
};
|
|
||||||
|
|
||||||
static PyGetSetDef lru_cache_getsetlist[] = {
|
|
||||||
{"__dict__", PyObject_GenericGetDict, PyObject_GenericSetDict},
|
|
||||||
{NULL}
|
|
||||||
};
|
|
||||||
|
|
||||||
static PyTypeObject lru_cache_type = {
|
|
||||||
PyVarObject_HEAD_INIT(NULL, 0)
|
|
||||||
"functools._lru_cache_wrapper", /* tp_name */
|
|
||||||
sizeof(lru_cache_object), /* tp_basicsize */
|
|
||||||
0, /* tp_itemsize */
|
|
||||||
/* methods */
|
|
||||||
(destructor)lru_cache_dealloc, /* tp_dealloc */
|
|
||||||
0, /* tp_print */
|
|
||||||
0, /* tp_getattr */
|
|
||||||
0, /* tp_setattr */
|
|
||||||
0, /* tp_reserved */
|
|
||||||
0, /* tp_repr */
|
|
||||||
0, /* tp_as_number */
|
|
||||||
0, /* tp_as_sequence */
|
|
||||||
0, /* tp_as_mapping */
|
|
||||||
0, /* tp_hash */
|
|
||||||
(ternaryfunc)lru_cache_call, /* tp_call */
|
|
||||||
0, /* tp_str */
|
|
||||||
0, /* tp_getattro */
|
|
||||||
0, /* tp_setattro */
|
|
||||||
0, /* tp_as_buffer */
|
|
||||||
Py_TPFLAGS_DEFAULT|Py_TPFLAGS_BASETYPE|Py_TPFLAGS_HAVE_GC,
|
|
||||||
/* tp_flags */
|
|
||||||
lru_cache_doc, /* tp_doc */
|
|
||||||
(traverseproc)lru_cache_tp_traverse,/* tp_traverse */
|
|
||||||
(inquiry)lru_cache_tp_clear, /* tp_clear */
|
|
||||||
0, /* tp_richcompare */
|
|
||||||
0, /* tp_weaklistoffset */
|
|
||||||
0, /* tp_iter */
|
|
||||||
0, /* tp_iternext */
|
|
||||||
lru_cache_methods, /* tp_methods */
|
|
||||||
0, /* tp_members */
|
|
||||||
lru_cache_getsetlist, /* tp_getset */
|
|
||||||
0, /* tp_base */
|
|
||||||
0, /* tp_dict */
|
|
||||||
0, /* tp_descr_get */
|
|
||||||
0, /* tp_descr_set */
|
|
||||||
offsetof(lru_cache_object, dict), /* tp_dictoffset */
|
|
||||||
0, /* tp_init */
|
|
||||||
0, /* tp_alloc */
|
|
||||||
lru_cache_new, /* tp_new */
|
|
||||||
};
|
|
||||||
|
|
||||||
/* module level code ********************************************************/
|
/* module level code ********************************************************/
|
||||||
|
|
||||||
PyDoc_STRVAR(module_doc,
|
PyDoc_STRVAR(module_doc,
|
||||||
|
@ -1135,11 +602,6 @@ static PyMethodDef module_methods[] = {
|
||||||
{NULL, NULL} /* sentinel */
|
{NULL, NULL} /* sentinel */
|
||||||
};
|
};
|
||||||
|
|
||||||
static void
|
|
||||||
module_free(void *m)
|
|
||||||
{
|
|
||||||
Py_CLEAR(kwd_mark);
|
|
||||||
}
|
|
||||||
|
|
||||||
static struct PyModuleDef _functoolsmodule = {
|
static struct PyModuleDef _functoolsmodule = {
|
||||||
PyModuleDef_HEAD_INIT,
|
PyModuleDef_HEAD_INIT,
|
||||||
|
@ -1150,7 +612,7 @@ static struct PyModuleDef _functoolsmodule = {
|
||||||
NULL,
|
NULL,
|
||||||
NULL,
|
NULL,
|
||||||
NULL,
|
NULL,
|
||||||
module_free,
|
NULL
|
||||||
};
|
};
|
||||||
|
|
||||||
PyMODINIT_FUNC
|
PyMODINIT_FUNC
|
||||||
|
@ -1161,7 +623,6 @@ PyInit__functools(void)
|
||||||
char *name;
|
char *name;
|
||||||
PyTypeObject *typelist[] = {
|
PyTypeObject *typelist[] = {
|
||||||
&partial_type,
|
&partial_type,
|
||||||
&lru_cache_type,
|
|
||||||
NULL
|
NULL
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -1169,12 +630,6 @@ PyInit__functools(void)
|
||||||
if (m == NULL)
|
if (m == NULL)
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
kwd_mark = PyObject_CallObject((PyObject *)&PyBaseObject_Type, NULL);
|
|
||||||
if (!kwd_mark) {
|
|
||||||
Py_DECREF(m);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
for (i=0 ; typelist[i] != NULL ; i++) {
|
for (i=0 ; typelist[i] != NULL ; i++) {
|
||||||
if (PyType_Ready(typelist[i]) < 0) {
|
if (PyType_Ready(typelist[i]) < 0) {
|
||||||
Py_DECREF(m);
|
Py_DECREF(m);
|
||||||
|
|
Loading…
Reference in New Issue