mirror of https://github.com/mahmoud/boltons.git
52 lines
1.6 KiB
ReStructuredText
52 lines
1.6 KiB
ReStructuredText
``cacheutils`` - Caches and caching
|
|
===================================
|
|
|
|
.. automodule:: boltons.cacheutils
|
|
|
|
|
|
Least-Recently Inserted (LRI)
|
|
-----------------------------
|
|
|
|
The :class:`LRI` is the simpler cache, implementing a very simple first-in,
|
|
first-out (FIFO) approach to cache eviction. If the use case calls for
|
|
simple, very-low overhead caching, such as somewhat expensive local
|
|
operations (e.g., string operations), then the LRI is likely the right
|
|
choice.
|
|
|
|
.. autoclass:: boltons.cacheutils.LRI
|
|
:members:
|
|
|
|
Least-Recently Used (LRU)
|
|
-------------------------
|
|
|
|
The :class:`LRU` is the more advanced cache, but it's still quite
|
|
simple. When it reaches capacity, a new insertion replaces the
|
|
least-recently used item. This strategy makes the LRU a more effective
|
|
cache than the LRI for a wide variety of applications, but also
|
|
entails more operations for all of its APIs, especially reads. Unlike
|
|
the :class:`LRI`, the LRU has threadsafety built in.
|
|
|
|
.. autoclass:: boltons.cacheutils.LRU
|
|
:members:
|
|
|
|
|
|
Automatic function caching
|
|
--------------------------
|
|
|
|
Continuing in the theme of cache tunability and experimentation,
|
|
``cacheutils`` also offers a way to pluggably cache function return
|
|
values: the :func:`cached` function decorator.
|
|
|
|
.. autofunction:: boltons.cacheutils.cached
|
|
|
|
Similar functionality can be found in Python 3.4's
|
|
:func:`functools.lru_cache` decorator, but the functools approach does
|
|
not support the same cache strategy modification or sharing the cache
|
|
object across multiple functions.
|
|
|
|
Threshold-bounded Counting
|
|
--------------------------
|
|
|
|
.. autoclass:: boltons.cacheutils.ThresholdCounter
|
|
:members:
|