summaryrefslogtreecommitdiffstats
path: root/Lib/functools.py
diff options
context:
space:
mode:
authorRaymond Hettinger <python@rcn.com>2013-03-09 04:14:46 (GMT)
committerRaymond Hettinger <python@rcn.com>2013-03-09 04:14:46 (GMT)
commita28c55a2e484ea81b4f659f987893be3638cb982 (patch)
tree252109e5d77ebf95330db61a6b49345aaa142939 /Lib/functools.py
parent11c08fd4b4ad84a3d6c4852e15afcbd9290aea99 (diff)
parentf96b2b03bbce6119b1993523a0dc2a9c020f0a90 (diff)
downloadcpython-a28c55a2e484ea81b4f659f987893be3638cb982.zip
cpython-a28c55a2e484ea81b4f659f987893be3638cb982.tar.gz
cpython-a28c55a2e484ea81b4f659f987893be3638cb982.tar.bz2
merge
Diffstat (limited to 'Lib/functools.py')
-rw-r--r--Lib/functools.py47
1 files changed, 35 insertions, 12 deletions
diff --git a/Lib/functools.py b/Lib/functools.py
index 9846a55..c436a52 100644
--- a/Lib/functools.py
+++ b/Lib/functools.py
@@ -20,7 +20,7 @@ try:
from _thread import RLock
except:
class RLock:
- 'Dummy reentrant lock'
+ 'Dummy reentrant lock for builds without threads'
def __enter__(self): pass
def __exit__(self, exctype, excinst, exctb): pass
@@ -172,6 +172,12 @@ except ImportError:
_CacheInfo = namedtuple("CacheInfo", ["hits", "misses", "maxsize", "currsize"])
class _HashedSeq(list):
+ """ This class guarantees that hash() will be called no more than once
+ per element. This is important because the lru_cache() will hash
+ the key multiple times on a cache miss.
+
+ """
+
__slots__ = 'hashvalue'
def __init__(self, tup, hash=hash):
@@ -185,7 +191,16 @@ def _make_key(args, kwds, typed,
kwd_mark = (object(),),
fasttypes = {int, str, frozenset, type(None)},
sorted=sorted, tuple=tuple, type=type, len=len):
- 'Make a cache key from optionally typed positional and keyword arguments'
+ """Make a cache key from optionally typed positional and keyword arguments
+
+ The key is constructed in a way that is flat as possible rather than
+ as a nested structure that would take more memory.
+
+ If there is only a single argument and its data type is known to cache
+ its hash value, then that argument is returned without a wrapper. This
+ saves space and improves lookup speed.
+
+ """
key = args
if kwds:
sorted_items = sorted(kwds.items())
@@ -242,7 +257,7 @@ def lru_cache(maxsize=128, typed=False):
if maxsize == 0:
def wrapper(*args, **kwds):
- # no caching, just a statistics update after a successful call
+ # No caching -- just a statistics update after a successful call
nonlocal misses
result = user_function(*args, **kwds)
misses += 1
@@ -272,8 +287,8 @@ def lru_cache(maxsize=128, typed=False):
with lock:
link = cache_get(key)
if link is not None:
- # move the link to the front of the circular queue
- link_prev, link_next, key, result = link
+ # Move the link to the front of the circular queue
+ link_prev, link_next, _key, result = link
link_prev[NEXT] = link_next
link_next[PREV] = link_prev
last = root[PREV]
@@ -285,26 +300,34 @@ def lru_cache(maxsize=128, typed=False):
result = user_function(*args, **kwds)
with lock:
if key in cache:
- # getting here means that this same key was added to the
- # cache while the lock was released. since the link
+ # Getting here means that this same key was added to the
+ # cache while the lock was released. Since the link
# update is already done, we need only return the
# computed result and update the count of misses.
pass
elif full:
- # use the old root to store the new key and result
+ # Use the old root to store the new key and result.
oldroot = root
oldroot[KEY] = key
oldroot[RESULT] = result
- # empty the oldest link and make it the new root
+ # Empty the oldest link and make it the new root.
+ # Keep a reference to the old key and old result to
+ # prevent their ref counts from going to zero during the
+ # update. That will prevent potentially arbitrary object
+ # clean-up code (i.e. __del__) from running while we're
+ # still adjusting the links.
root = oldroot[NEXT]
oldkey = root[KEY]
- oldvalue = root[RESULT]
+ oldresult = root[RESULT]
root[KEY] = root[RESULT] = None
- # now update the cache dictionary for the new links
+ # Now update the cache dictionary.
del cache[oldkey]
+ # Save the potentially reentrant cache[key] assignment
+ # for last, after the root and links have been put in
+ # a consistent state.
cache[key] = oldroot
else:
- # put result in a new link at the front of the queue
+ # Put result in a new link at the front of the queue.
last = root[PREV]
link = [last, root, key, result]
last[NEXT] = root[PREV] = cache[key] = link