Django’s cache framework

A fundamental trade-off in dynamic websites is, well, they’re dynamic. Eachtime a user requests a page, the Web server makes all sorts of calculations –from database queries to template rendering to business logic – to create thepage that your site’s visitor sees. This is a lot more expensive, from aprocessing-overhead perspective, than your standardread-a-file-off-the-filesystem server arrangement.

For most Web applications, this overhead isn’t a big deal. Most Webapplications aren’t washingtonpost.com or slashdot.org; they’re small-to medium-sized sites with so-so traffic. But for medium- to high-trafficsites, it’s essential to cut as much overhead as possible.

That’s where caching comes in.

To cache something is to save the result of an expensive calculation so thatyou don’t have to perform the calculation next time. Here’s some pseudocodeexplaining how this would work for a dynamically generated Web page:

  1. given a URL, try finding that page in the cache
  2. if the page is in the cache:
  3. return the cached page
  4. else:
  5. generate the page
  6. save the generated page in the cache (for next time)
  7. return the generated page

Django comes with a robust cache system that lets you save dynamic pages sothey don’t have to be calculated for each request. For convenience, Djangooffers different levels of cache granularity: You can cache the output ofspecific views, you can cache only the pieces that are difficult to produce,or you can cache your entire site.

Django also works well with “downstream” caches, such as Squid and browser-based caches. These are the types ofcaches that you don’t directly control but to which you can provide hints (viaHTTP headers) about which parts of your site should be cached, and how.

See also

The Cache Framework design philosophyexplains a few of the design decisions of the framework.

Setting up the cache

The cache system requires a small amount of setup. Namely, you have to tell itwhere your cached data should live – whether in a database, on the filesystemor directly in memory. This is an important decision that affects your cache’sperformance; yes, some cache types are faster than others.

Your cache preference goes in the CACHES setting in yoursettings file. Here’s an explanation of all available values forCACHES.

Memcached

The fastest, most efficient type of cache supported natively by Django,Memcached is an entirely memory-based cache server, originally developedto handle high loads at LiveJournal.com and subsequently open-sourced byDanga Interactive. It is used by sites such as Facebook and Wikipedia toreduce database access and dramatically increase site performance.

Memcached runs as a daemon and is allotted a specified amount of RAM. All itdoes is provide a fast interface for adding, retrieving and deleting data inthe cache. All data is stored directly in memory, so there’s no overhead ofdatabase or filesystem usage.

After installing Memcached itself, you’ll need to install a Memcachedbinding. There are several Python Memcached bindings available; thetwo most common are python-memcached and pylibmc.

To use Memcached with Django:

  • Set BACKEND todjango.core.cache.backends.memcached.MemcachedCache ordjango.core.cache.backends.memcached.PyLibMCCache (dependingon your chosen memcached binding)
  • Set LOCATION to ip:port values,where ip is the IP address of the Memcached daemon and port is theport on which Memcached is running, or to a unix:path value, wherepath is the path to a Memcached Unix socket file.In this example, Memcached is running on localhost (127.0.0.1) port 11211, usingthe python-memcached binding:
  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'django.core.cache.backends.memcached.MemcachedCache',
  4. 'LOCATION': '127.0.0.1:11211',
  5. }
  6. }

In this example, Memcached is available through a local Unix socket file/tmp/memcached.sock using the python-memcached binding:

  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'django.core.cache.backends.memcached.MemcachedCache',
  4. 'LOCATION': 'unix:/tmp/memcached.sock',
  5. }
  6. }

When using the pylibmc binding, do not include the unix:/ prefix:

  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'django.core.cache.backends.memcached.PyLibMCCache',
  4. 'LOCATION': '/tmp/memcached.sock',
  5. }
  6. }

One excellent feature of Memcached is its ability to share a cache overmultiple servers. This means you can run Memcached daemons on multiplemachines, and the program will treat the group of machines as a _single_cache, without the need to duplicate cache values on each machine. To takeadvantage of this feature, include all server addresses inLOCATION, either as a semicolon or commadelimited string, or as a list.

In this example, the cache is shared over Memcached instances running on IPaddress 172.19.26.240 and 172.19.26.242, both on port 11211:

  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'django.core.cache.backends.memcached.MemcachedCache',
  4. 'LOCATION': [
  5. '172.19.26.240:11211',
  6. '172.19.26.242:11211',
  7. ]
  8. }
  9. }

In the following example, the cache is shared over Memcached instances runningon the IP addresses 172.19.26.240 (port 11211), 172.19.26.242 (port 11212), and172.19.26.244 (port 11213):

  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'django.core.cache.backends.memcached.MemcachedCache',
  4. 'LOCATION': [
  5. '172.19.26.240:11211',
  6. '172.19.26.242:11212',
  7. '172.19.26.244:11213',
  8. ]
  9. }
  10. }

A final point about Memcached is that memory-based caching has adisadvantage: because the cached data is stored in memory, the data will belost if your server crashes. Clearly, memory isn’t intended for permanent datastorage, so don’t rely on memory-based caching as your only data storage.Without a doubt, none of the Django caching backends should be used forpermanent storage – they’re all intended to be solutions for caching, notstorage – but we point this out here because memory-based caching isparticularly temporary.

Database caching

Django can store its cached data in your database. This works best if you’vegot a fast, well-indexed database server.

To use a database table as your cache backend:

  • Set BACKEND todjango.core.cache.backends.db.DatabaseCache
  • Set LOCATION to tablename, the name of thedatabase table. This name can be whatever you want, as long as it’s a validtable name that’s not already being used in your database.In this example, the cache table’s name is my_cache_table:
  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'django.core.cache.backends.db.DatabaseCache',
  4. 'LOCATION': 'my_cache_table',
  5. }
  6. }

Creating the cache table

Before using the database cache, you must create the cache table with thiscommand:

  1. python manage.py createcachetable

This creates a table in your database that is in the proper format thatDjango’s database-cache system expects. The name of the table is taken fromLOCATION.

If you are using multiple database caches, createcachetable createsone table for each cache.

If you are using multiple databases, createcachetable observes theallow_migrate() method of your database routers (see below).

Like migrate, createcachetable won’t touch an existingtable. It will only create missing tables.

To print the SQL that would be run, rather than run it, use thecreatecachetable —dry-run option.

Multiple databases

If you use database caching with multiple databases, you’ll also needto set up routing instructions for your database cache table. For thepurposes of routing, the database cache table appears as a model namedCacheEntry, in an application named django_cache. This modelwon’t appear in the models cache, but the model details can be usedfor routing purposes.

For example, the following router would direct all cache readoperations to cache_replica, and all write operations tocache_primary. The cache table will only be synchronized ontocache_primary:

  1. class CacheRouter:
  2. """A router to control all database cache operations"""
  3.  
  4. def db_for_read(self, model, **hints):
  5. "All cache read operations go to the replica"
  6. if model._meta.app_label == 'django_cache':
  7. return 'cache_replica'
  8. return None
  9.  
  10. def db_for_write(self, model, **hints):
  11. "All cache write operations go to primary"
  12. if model._meta.app_label == 'django_cache':
  13. return 'cache_primary'
  14. return None
  15.  
  16. def allow_migrate(self, db, app_label, model_name=None, **hints):
  17. "Only install the cache model on primary"
  18. if app_label == 'django_cache':
  19. return db == 'cache_primary'
  20. return None

If you don’t specify routing directions for the database cache model,the cache backend will use the default database.

Of course, if you don’t use the database cache backend, you don’t needto worry about providing routing instructions for the database cachemodel.

Filesystem caching

The file-based backend serializes and stores each cache value as a separatefile. To use this backend set BACKEND to"django.core.cache.backends.filebased.FileBasedCache" andLOCATION to a suitable directory. For example,to store cached data in /var/tmp/django_cache, use this setting:

  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache',
  4. 'LOCATION': '/var/tmp/django_cache',
  5. }
  6. }

If you’re on Windows, put the drive letter at the beginning of the path,like this:

  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache',
  4. 'LOCATION': 'c:/foo/bar',
  5. }
  6. }

The directory path should be absolute – that is, it should start at the rootof your filesystem. It doesn’t matter whether you put a slash at the end of thesetting.

Make sure the directory pointed-to by this setting exists and is readable andwritable by the system user under which your Web server runs. Continuing theabove example, if your server runs as the user apache, make sure thedirectory /var/tmp/django_cache exists and is readable and writable by theuser apache.

Local-memory caching

This is the default cache if another is not specified in your settings file. Ifyou want the speed advantages of in-memory caching but don’t have the capabilityof running Memcached, consider the local-memory cache backend. This cache isper-process (see below) and thread-safe. To use it, set BACKEND to "django.core.cache.backends.locmem.LocMemCache". Forexample:

  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'django.core.cache.backends.locmem.LocMemCache',
  4. 'LOCATION': 'unique-snowflake',
  5. }
  6. }

The cache LOCATION is used to identify individualmemory stores. If you only have one locmem cache, you can omit theLOCATION; however, if you have more than one localmemory cache, you will need to assign a name to at least one of them inorder to keep them separate.

The cache uses a least-recently-used (LRU) culling strategy.

Note that each process will have its own private cache instance, which means nocross-process caching is possible. This obviously also means the local memorycache isn’t particularly memory-efficient, so it’s probably not a good choicefor production environments. It’s nice for development.

Dummy caching (for development)

Finally, Django comes with a “dummy” cache that doesn’t actually cache – itjust implements the cache interface without doing anything.

This is useful if you have a production site that uses heavy-duty caching invarious places but a development/test environment where you don’t want to cacheand don’t want to have to change your code to special-case the latter. Toactivate dummy caching, set BACKEND like so:

  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'django.core.cache.backends.dummy.DummyCache',
  4. }
  5. }

Using a custom cache backend

While Django includes support for a number of cache backends out-of-the-box,sometimes you might want to use a customized cache backend. To use an externalcache backend with Django, use the Python import path as theBACKEND of the CACHES setting, like so:

  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'path.to.backend',
  4. }
  5. }

If you’re building your own backend, you can use the standard cache backendsas reference implementations. You’ll find the code in thedjango/core/cache/backends/ directory of the Django source.

Note: Without a really compelling reason, such as a host that doesn’t supportthem, you should stick to the cache backends included with Django. They’vebeen well-tested and are well-documented.

Cache arguments

Each cache backend can be given additional arguments to control cachingbehavior. These arguments are provided as additional keys in theCACHES setting. Valid arguments are as follows:

  • TIMEOUT: The default timeout, inseconds, to use for the cache. This argument defaults to 300 seconds (5 minutes).You can set TIMEOUT to None so that, by default, cache keys neverexpire. A value of 0 causes keys to immediately expire (effectively“don’t cache”).

  • OPTIONS: Any options that should bepassed to the cache backend. The list of valid options will varywith each backend, and cache backends backed by a third-party librarywill pass their options directly to the underlying cache library.

Cache backends that implement their own culling strategy (i.e.,the locmem, filesystem and database backends) willhonor the following options:

  • MAX_ENTRIES: The maximum number of entries allowed inthe cache before old values are deleted. This argumentdefaults to 300.

  • CULL_FREQUENCY: The fraction of entries that are culledwhen MAX_ENTRIES is reached. The actual ratio is1 / CULL_FREQUENCY, so set CULL_FREQUENCY to 2 tocull half the entries when MAX_ENTRIES is reached. This argumentshould be an integer and defaults to 3.

A value of 0 for CULL_FREQUENCY means that theentire cache will be dumped when MAX_ENTRIES is reached.On some backends (database in particular) this makes culling _much_faster at the expense of more cache misses.

Memcached backends pass the contents of OPTIONSas keyword arguments to the client constructors, allowing for more advancedcontrol of client behavior. For example usage, see below.

  • KEY_PREFIX: A string that will beautomatically included (prepended by default) to all cache keysused by the Django server.

See the cache documentation formore information.

  • VERSION: The default version numberfor cache keys generated by the Django server.

See the cache documentation for moreinformation.

  • KEY_FUNCTIONA string containing a dotted path to a function that defines howto compose a prefix, version and key into a final cache key.

See the cache documentationfor more information.

In this example, a filesystem backend is being configured with a timeoutof 60 seconds, and a maximum capacity of 1000 items:

  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache',
  4. 'LOCATION': '/var/tmp/django_cache',
  5. 'TIMEOUT': 60,
  6. 'OPTIONS': {
  7. 'MAX_ENTRIES': 1000
  8. }
  9. }
  10. }

Here’s an example configuration for a python-memcached based backend withan object size limit of 2MB:

  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'django.core.cache.backends.memcached.MemcachedCache',
  4. 'LOCATION': '127.0.0.1:11211',
  5. 'OPTIONS': {
  6. 'server_max_value_length': 1024 * 1024 * 2,
  7. }
  8. }
  9. }

Here’s an example configuration for a pylibmc based backend that enablesthe binary protocol, SASL authentication, and the ketama behavior mode:

  1. CACHES = {
  2. 'default': {
  3. 'BACKEND': 'django.core.cache.backends.memcached.PyLibMCCache',
  4. 'LOCATION': '127.0.0.1:11211',
  5. 'OPTIONS': {
  6. 'binary': True,
  7. 'username': 'user',
  8. 'password': 'pass',
  9. 'behaviors': {
  10. 'ketama': True,
  11. }
  12. }
  13. }
  14. }

The per-site cache

Once the cache is set up, the simplest way to use caching is to cache yourentire site. You’ll need to add'django.middleware.cache.UpdateCacheMiddleware' and'django.middleware.cache.FetchFromCacheMiddleware' to yourMIDDLEWARE setting, as in this example:

  1. MIDDLEWARE = [
  2. 'django.middleware.cache.UpdateCacheMiddleware',
  3. 'django.middleware.common.CommonMiddleware',
  4. 'django.middleware.cache.FetchFromCacheMiddleware',
  5. ]

Note

No, that’s not a typo: the “update” middleware must be first in the list,and the “fetch” middleware must be last. The details are a bit obscure, butsee Order of MIDDLEWARE below if you’d like the full story.

Then, add the following required settings to your Django settings file:

  • CACHE_MIDDLEWARE_ALIAS – The cache alias to use for storage.
  • CACHE_MIDDLEWARE_SECONDS – The number of seconds each page shouldbe cached.
  • CACHE_MIDDLEWARE_KEY_PREFIX – If the cache is shared acrossmultiple sites using the same Django installation, set this to the name ofthe site, or some other string that is unique to this Django instance, toprevent key collisions. Use an empty string if you don’t care.FetchFromCacheMiddleware caches GET and HEAD responses with status 200,where the request and response headers allow. Responses to requests for the sameURL with different query parameters are considered to be unique pages and arecached separately. This middleware expects that a HEAD request is answered withthe same response headers as the corresponding GET request; in which case it canreturn a cached GET response for HEAD request.

Additionally, UpdateCacheMiddleware automatically sets a few headers in eachHttpResponse:

If a view sets its own cache expiry time (i.e. it has a max-age section inits Cache-Control header) then the page will be cached until the expirytime, rather than CACHE_MIDDLEWARE_SECONDS. Using the decorators indjango.views.decorators.cache you can easily set a view’s expiry time(using the cache_control() decorator) ordisable caching for a view (using thenever_cache() decorator). See theusing other headers section for more on these decorators.

If USE_I18N is set to True then the generated cache key willinclude the name of the active language – see alsoHow Django discovers language preference). This allows you to easilycache multilingual sites without having to create the cache key yourself.

Cache keys also include the active language whenUSE_L10N is set to True and the current time zone when USE_TZ is set to True.

The per-view cache

  • django.views.decorators.cache.cache_page()
  • A more granular way to use the caching framework is by caching the output ofindividual views. django.views.decorators.cache defines a cache_pagedecorator that will automatically cache the view’s response for you:
  1. from django.views.decorators.cache import cache_page
  2.  
  3. @cache_page(60 * 15)
  4. def my_view(request):
  5. ...

cache_page takes a single argument: the cache timeout, in seconds. In theabove example, the result of the my_view() view will be cached for 15minutes. (Note that we’ve written it as 60 15 for the purpose ofreadability. 60 15 will be evaluated to 900 – that is, 15 minutesmultiplied by 60 seconds per minute.)

The per-view cache, like the per-site cache, is keyed off of the URL. Ifmultiple URLs point at the same view, each URL will be cached separately.Continuing the my_view example, if your URLconf looks like this:

  1. urlpatterns = [
  2. path('foo/<int:code>/', my_view),
  3. ]

then requests to /foo/1/ and /foo/23/ will be cached separately, asyou may expect. But once a particular URL (e.g., /foo/23/) has beenrequested, subsequent requests to that URL will use the cache.

cache_page can also take an optional keyword argument, cache,which directs the decorator to use a specific cache (from yourCACHES setting) when caching view results. By default, thedefault cache will be used, but you can specify any cache youwant:

  1. @cache_page(60 * 15, cache="special_cache")def my_view(request):

You can also override the cache prefix on a per-view basis. cache_pagetakes an optional keyword argument, key_prefix,which works in the same way as the CACHE_MIDDLEWARE_KEY_PREFIXsetting for the middleware. It can be used like this:

  1. @cache_page(60 * 15, key_prefix="site1")def my_view(request):

The key_prefix and cache arguments may be specified together. Thekey_prefix argument and the KEY_PREFIXspecified under CACHES will be concatenated.

Specifying per-view cache in the URLconf

The examples in the previous section have hard-coded the fact that the view iscached, because cache_page alters the my_view function in place. Thisapproach couples your view to the cache system, which is not ideal for severalreasons. For instance, you might want to reuse the view functions on another,cache-less site, or you might want to distribute the views to people who mightwant to use them without being cached. The solution to these problems is tospecify the per-view cache in the URLconf rather than next to the view functionsthemselves.

You can do so by wrapping the view function with cache_page when you referto it in the URLconf. Here’s the old URLconf from earlier:

  1. urlpatterns = [
  2. path('foo/<int:code>/', my_view),
  3. ]

Here’s the same thing, with my_view wrapped in cache_page:

  1. from django.views.decorators.cache import cache_page
  2.  
  3. urlpatterns = [
  4. path('foo/<int:code>/', cache_page(60 * 15)(my_view)),
  5. ]

Template fragment caching

If you’re after even more control, you can also cache template fragments usingthe cache template tag. To give your template access to this tag, put{% load cache %} near the top of your template.

The {% cache %} template tag caches the contents of the block for a givenamount of time. It takes at least two arguments: the cache timeout, in seconds,and the name to give the cache fragment. The fragment is cached forever iftimeout is None. The name will be taken as is, do not use a variable. Forexample:

  1. {% load cache %}
  2. {% cache 500 sidebar %}
  3. .. sidebar ..
  4. {% endcache %}

Sometimes you might want to cache multiple copies of a fragment depending onsome dynamic data that appears inside the fragment. For example, you might want aseparate cached copy of the sidebar used in the previous example for every userof your site. Do this by passing one or more additional arguments, which may bevariables with or without filters, to the {% cache %} template tag touniquely identify the cache fragment:

  1. {% load cache %}
  2. {% cache 500 sidebar request.user.username %}
  3. .. sidebar for logged in user ..
  4. {% endcache %}

If USE_I18N is set to True the per-site middleware cache willrespect the active language. For the cache templatetag you could use one of thetranslation-specific variables available intemplates to achieve the same result:

  1. {% load i18n %}
  2. {% load cache %}
  3.  
  4. {% get_current_language as LANGUAGE_CODE %}
  5.  
  6. {% cache 600 welcome LANGUAGE_CODE %}
  7. {% trans "Welcome to example.com" %}
  8. {% endcache %}

The cache timeout can be a template variable, as long as the template variableresolves to an integer value. For example, if the template variablemy_timeout is set to the value 600, then the following two examples areequivalent:

  1. {% cache 600 sidebar %} ... {% endcache %}
  2. {% cache my_timeout sidebar %} ... {% endcache %}

This feature is useful in avoiding repetition in templates. You can set thetimeout in a variable, in one place, and reuse that value.

By default, the cache tag will try to use the cache called “template_fragments”.If no such cache exists, it will fall back to using the default cache. You mayselect an alternate cache backend to use with the using keyword argument,which must be the last argument to the tag.

  1. {% cache 300 local-thing ... using="localcache" %}

It is considered an error to specify a cache name that is not configured.

  • django.core.cache.utils.maketemplate_fragment_key(_fragment_name, vary_on=None)
  • If you want to obtain the cache key used for a cached fragment, you can usemake_template_fragment_key. fragment_name is the same as second argumentto the cache template tag; vary_on is a list of all additional argumentspassed to the tag. This function can be useful for invalidating or overwritinga cached item, for example:
  1. >>> from django.core.cache import cache
  2. >>> from django.core.cache.utils import make_template_fragment_key
  3. # cache key for {% cache 500 sidebar username %}
  4. >>> key = make_template_fragment_key('sidebar', [username])
  5. >>> cache.delete(key) # invalidates cached template fragment

The low-level cache API

Sometimes, caching an entire rendered page doesn’t gain you very much and is,in fact, inconvenient overkill.

Perhaps, for instance, your site includes a view whose results depend onseveral expensive queries, the results of which change at different intervals.In this case, it would not be ideal to use the full-page caching that theper-site or per-view cache strategies offer, because you wouldn’t want tocache the entire result (since some of the data changes often), but you’d stillwant to cache the results that rarely change.

For cases like this, Django exposes a low-level cache API. You can use this APIto store objects in the cache with any level of granularity you like. You cancache any Python object that can be pickled safely: strings, dictionaries,lists of model objects, and so forth. (Most common Python objects can bepickled; refer to the Python documentation for more information aboutpickling.)

Accessing the cache

  • django.core.cache.caches
  • You can access the caches configured in the CACHES settingthrough a dict-like object: django.core.cache.caches. Repeatedrequests for the same alias in the same thread will return the sameobject.
  1. >>> from django.core.cache import caches
  2. >>> cache1 = caches['myalias']
  3. >>> cache2 = caches['myalias']
  4. >>> cache1 is cache2
  5. True

If the named key does not exist, InvalidCacheBackendError will beraised.

To provide thread-safety, a different instance of the cache backend willbe returned for each thread.

  • django.core.cache.cache
  • As a shortcut, the default cache is available asdjango.core.cache.cache:
  1. >>> from django.core.cache import cache

This object is equivalent to caches['default'].

Basic usage

The basic interface is:

  • cache.set(key, value, timeout=DEFAULT_TIMEOUT, version=None)
  1. >>> cache.set('my_key', 'hello, world!', 30)
  • cache.get(key, default=None, version=None)
  1. >>> cache.get('my_key')
  2. 'hello, world!'

key should be a str, and value can be any picklable Python object.

The timeout argument is optional and defaults to the timeout argumentof the appropriate backend in the CACHES setting (explained above).It’s the number of seconds the value should be stored in the cache. Passing inNone for timeout will cache the value forever. A timeout of 0won’t cache the value.

If the object doesn’t exist in the cache, cache.get() returns None:

  1. >>> # Wait 30 seconds for 'my_key' to expire...
  2. >>> cache.get('my_key')
  3. None

We advise against storing the literal value None in the cache, because youwon’t be able to distinguish between your stored None value and a cachemiss signified by a return value of None.

cache.get() can take a default argument. This specifies which value toreturn if the object doesn’t exist in the cache:

  1. >>> cache.get('my_key', 'has expired')
  2. 'has expired'
  • cache.add(key, value, timeout=DEFAULT_TIMEOUT, version=None)
  • To add a key only if it doesn’t already exist, use the add() method.It takes the same parameters as set(), but it will not attempt toupdate the cache if the key specified is already present:
  1. >>> cache.set('add_key', 'Initial value')
  2. >>> cache.add('add_key', 'New value')
  3. >>> cache.get('add_key')
  4. 'Initial value'

If you need to know whether add() stored a value in the cache, you cancheck the return value. It will return True if the value was stored,False otherwise.

  • cache.getor_set(_key, default, timeout=DEFAULT_TIMEOUT, version=None)
  • If you want to get a key’s value or set a value if the key isn’t in the cache,there is the get_or_set() method. It takes the same parameters as get()but the default is set as the new cache value for that key, rather thanreturned:
  1. >>> cache.get('my_new_key') # returns None
  2. >>> cache.get_or_set('my_new_key', 'my new value', 100)
  3. 'my new value'

You can also pass any callable as a default value:

  1. >>> import datetime
  2. >>> cache.get_or_set('some-timestamp-key', datetime.datetime.now)
  3. datetime.datetime(2014, 12, 11, 0, 15, 49, 457920)
  • cache.getmany(_keys, version=None)
  • There’s also a get_many() interface that only hits the cache once.get_many() returns a dictionary with all the keys you asked for thatactually exist in the cache (and haven’t expired):
  1. >>> cache.set('a', 1)
  2. >>> cache.set('b', 2)
  3. >>> cache.set('c', 3)
  4. >>> cache.get_many(['a', 'b', 'c'])
  5. {'a': 1, 'b': 2, 'c': 3}
  • cache.setmany(_dict, timeout)
  • To set multiple values more efficiently, use set_many() to pass a dictionaryof key-value pairs:
  1. >>> cache.set_many({'a': 1, 'b': 2, 'c': 3})
  2. >>> cache.get_many(['a', 'b', 'c'])
  3. {'a': 1, 'b': 2, 'c': 3}

Like cache.set(), set_many() takes an optional timeout parameter.

On supported backends (memcached), set_many() returns a list of keys thatfailed to be inserted.

  • cache.delete(key, version=None)
  • You can delete keys explicitly with delete() to clear the cache for aparticular object:
  1. >>> cache.delete('a')
  • cache.deletemany(_keys, version=None)
  • If you want to clear a bunch of keys at once, delete_many() can take a listof keys to be cleared:
  1. >>> cache.delete_many(['a', 'b', 'c'])
  • cache.clear()
  • Finally, if you want to delete all the keys in the cache, usecache.clear(). Be careful with this; clear() will remove _everything_from the cache, not just the keys set by your application.
  1. >>> cache.clear()
  • cache.touch(key, timeout=DEFAULT_TIMEOUT, version=None)
  • cache.touch() sets a new expiration for a key. For example, to update a keyto expire 10 seconds from now:
  1. >>> cache.touch('a', 10)
  2. True

Like other methods, the timeout argument is optional and defaults to theTIMEOUT option of the appropriate backend in the CACHES setting.

touch() returns True if the key was successfully touched, Falseotherwise.

  • cache.incr(key, delta=1, version=None)
  • cache.decr(key, delta=1, version=None)
  • You can also increment or decrement a key that already exists using theincr() or decr() methods, respectively. By default, the existing cachevalue will be incremented or decremented by 1. Other increment/decrement valuescan be specified by providing an argument to the increment/decrement call. AValueError will be raised if you attempt to increment or decrement anonexistent cache key.:
  1. >>> cache.set('num', 1)
  2. >>> cache.incr('num')
  3. 2
  4. >>> cache.incr('num', 10)
  5. 12
  6. >>> cache.decr('num')
  7. 11
  8. >>> cache.decr('num', 5)
  9. 6

Note

incr()/decr() methods are not guaranteed to be atomic. On thosebackends that support atomic increment/decrement (most notably, thememcached backend), increment and decrement operations will be atomic.However, if the backend doesn’t natively provide an increment/decrementoperation, it will be implemented using a two-step retrieve/update.

  • cache.close()
  • You can close the connection to your cache with close() if implemented bythe cache backend.
  1. >>> cache.close()

Note

For caches that don’t implement close methods it is a no-op.

Cache key prefixing

If you are sharing a cache instance between servers, or between yourproduction and development environments, it’s possible for data cachedby one server to be used by another server. If the format of cacheddata is different between servers, this can lead to some very hard todiagnose problems.

To prevent this, Django provides the ability to prefix all cache keysused by a server. When a particular cache key is saved or retrieved,Django will automatically prefix the cache key with the value of theKEY_PREFIX cache setting.

By ensuring each Django instance has a differentKEY_PREFIX, you can ensure that there will be nocollisions in cache values.

Cache versioning

When you change running code that uses cached values, you may need topurge any existing cached values. The easiest way to do this is toflush the entire cache, but this can lead to the loss of cache valuesthat are still valid and useful.

Django provides a better way to target individual cache values.Django’s cache framework has a system-wide version identifier,specified using the VERSION cache setting.The value of this setting is automatically combined with the cacheprefix and the user-provided cache key to obtain the final cache key.

By default, any key request will automatically include the sitedefault cache key version. However, the primitive cache functions allinclude a version argument, so you can specify a particular cachekey version to set or get. For example:

  1. >>> # Set version 2 of a cache key
  2. >>> cache.set('my_key', 'hello world!', version=2)
  3. >>> # Get the default version (assuming version=1)
  4. >>> cache.get('my_key')
  5. None
  6. >>> # Get version 2 of the same key
  7. >>> cache.get('my_key', version=2)
  8. 'hello world!'

The version of a specific key can be incremented and decremented usingthe incr_version() and decr_version() methods. Thisenables specific keys to be bumped to a new version, leaving otherkeys unaffected. Continuing our previous example:

  1. >>> # Increment the version of 'my_key'
  2. >>> cache.incr_version('my_key')
  3. >>> # The default version still isn't available
  4. >>> cache.get('my_key')
  5. None
  6. # Version 2 isn't available, either
  7. >>> cache.get('my_key', version=2)
  8. None
  9. >>> # But version 3 *is* available
  10. >>> cache.get('my_key', version=3)
  11. 'hello world!'

Cache key transformation

As described in the previous two sections, the cache key provided by auser is not used verbatim – it is combined with the cache prefix andkey version to provide a final cache key. By default, the three partsare joined using colons to produce a final string:

  1. def make_key(key, key_prefix, version):
  2. return ':'.join([key_prefix, str(version), key])

If you want to combine the parts in different ways, or apply otherprocessing to the final key (e.g., taking a hash digest of the keyparts), you can provide a custom key function.

The KEY_FUNCTION cache settingspecifies a dotted-path to a function matching the prototype ofmake_key() above. If provided, this custom key function willbe used instead of the default key combining function.

Cache key warnings

Memcached, the most commonly-used production cache backend, does not allowcache keys longer than 250 characters or containing whitespace or controlcharacters, and using such keys will cause an exception. To encouragecache-portable code and minimize unpleasant surprises, the other built-in cachebackends issue a warning (django.core.cache.backends.base.CacheKeyWarning)if a key is used that would cause an error on memcached.

If you are using a production backend that can accept a wider range of keys (acustom backend, or one of the non-memcached built-in backends), and want to usethis wider range without warnings, you can silence CacheKeyWarning withthis code in the management module of one of yourINSTALLED_APPS:

  1. import warnings
  2.  
  3. from django.core.cache import CacheKeyWarning
  4.  
  5. warnings.simplefilter("ignore", CacheKeyWarning)

If you want to instead provide custom key validation logic for one of thebuilt-in backends, you can subclass it, override just the validate_keymethod, and follow the instructions for using a custom cache backend. Forinstance, to do this for the locmem backend, put this code in a module:

  1. from django.core.cache.backends.locmem import LocMemCache
  2.  
  3. class CustomLocMemCache(LocMemCache):
  4. def validate_key(self, key):
  5. """Custom validation, raising exceptions or warnings as needed."""
  6. ...

…and use the dotted Python path to this class in theBACKEND portion of your CACHES setting.

Downstream caches

So far, this document has focused on caching your own data. But another typeof caching is relevant to Web development, too: caching performed by“downstream” caches. These are systems that cache pages for users even beforethe request reaches your website.

Here are a few examples of downstream caches:

  • Your ISP may cache certain pages, so if you requested a page fromhttps://example.com/, your ISP would send you the page without having toaccess example.com directly. The maintainers of example.com have noknowledge of this caching; the ISP sits between example.com and your Webbrowser, handling all of the caching transparently.
  • Your Django website may sit behind a proxy cache, such as Squid WebProxy Cache (http://www.squid-cache.org/), that caches pages forperformance. In this case, each request first would be handled by theproxy, and it would be passed to your application only if needed.
  • Your Web browser caches pages, too. If a Web page sends out theappropriate headers, your browser will use the local cached copy forsubsequent requests to that page, without even contacting the Web pageagain to see whether it has changed.Downstream caching is a nice efficiency boost, but there’s a danger to it:Many Web pages’ contents differ based on authentication and a host of othervariables, and cache systems that blindly save pages based purely on URLs couldexpose incorrect or sensitive data to subsequent visitors to those pages.

For example, if you operate a Web email system, then the contents of the“inbox” page depend on which user is logged in. If an ISP blindly cached yoursite, then the first user who logged in through that ISP would have theiruser-specific inbox page cached for subsequent visitors to the site. That’snot cool.

Fortunately, HTTP provides a solution to this problem. A number of HTTP headersexist to instruct downstream caches to differ their cache contents depending ondesignated variables, and to tell caching mechanisms not to cache particularpages. We’ll look at some of these headers in the sections that follow.

Using Vary headers

The Vary header defines which request headers a cachemechanism should take into account when building its cache key. For example, ifthe contents of a Web page depend on a user’s language preference, the page issaid to “vary on language.”

By default, Django’s cache system creates its cache keys using the requestedfully-qualified URL – e.g.,"https://www.example.com/stories/2005/?order_by=author&#34;. This means everyrequest to that URL will use the same cached version, regardless of user-agentdifferences such as cookies or language preferences. However, if this pageproduces different content based on some difference in request headers – suchas a cookie, or a language, or a user-agent – you’ll need to use the Varyheader to tell caching mechanisms that the page output depends on those things.

To do this in Django, use the convenientdjango.views.decorators.vary.vary_on_headers() view decorator, like so:

  1. from django.views.decorators.vary import vary_on_headers
  2.  
  3. @vary_on_headers('User-Agent')
  4. def my_view(request):
  5. ...

In this case, a caching mechanism (such as Django’s own cache middleware) willcache a separate version of the page for each unique user-agent.

The advantage to using the varyon_headers decorator rather than manuallysetting the Vary header (using something likeresponse['Vary'] = 'user-agent') is that the decorator _adds to theVary header (which may already exist), rather than setting it from scratchand potentially overriding anything that was already in there.

You can pass multiple headers to vary_on_headers():

  1. @vary_on_headers('User-Agent', 'Cookie')def my_view(request):

This tells downstream caches to vary on both, which means each combination ofuser-agent and cookie will get its own cache value. For example, a request withthe user-agent Mozilla and the cookie value foo=bar will be considereddifferent from a request with the user-agent Mozilla and the cookie valuefoo=ham.

Because varying on cookie is so common, there’s adjango.views.decorators.vary.vary_on_cookie() decorator. These two viewsare equivalent:

  1. @vary_on_cookiedef my_view(request):

  2. @vary_on_headers('Cookie')def my_view(request):

The headers you pass to vary_on_headers are not case sensitive;"User-Agent" is the same thing as "user-agent".

You can also use a helper function, django.utils.cache.patch_vary_headers(),directly. This function sets, or adds to, the Vary header. For example:

  1. from django.shortcuts import render
  2. from django.utils.cache import patch_vary_headers
  3.  
  4. def my_view(request):
  5. ...
  6. response = render(request, 'template_name', context)
  7. patch_vary_headers(response, ['Cookie'])
  8. return response

patch_vary_headers takes an HttpResponse instance asits first argument and a list/tuple of case-insensitive header names as itssecond argument.

For more on Vary headers, see the official Vary spec.

Controlling cache: Using other headers

Other problems with caching are the privacy of data and the question of wheredata should be stored in a cascade of caches.

A user usually faces two kinds of caches: their own browser cache (a privatecache) and their provider’s cache (a public cache). A public cache is used bymultiple users and controlled by someone else. This poses problems withsensitive data–you don’t want, say, your bank account number stored in apublic cache. So Web applications need a way to tell caches which data isprivate and which is public.

The solution is to indicate a page’s cache should be “private.” To do this inDjango, use the cache_control() viewdecorator. Example:

  1. from django.views.decorators.cache import cache_control
  2.  
  3. @cache_control(private=True)
  4. def my_view(request):
  5. ...

This decorator takes care of sending out the appropriate HTTP header behind thescenes.

Note that the cache control settings “private” and “public” are mutuallyexclusive. The decorator ensures that the “public” directive is removed if“private” should be set (and vice versa). An example use of the two directiveswould be a blog site that offers both private and public entries. Publicentries may be cached on any shared cache. The following code usespatch_cache_control(), the manual way to modify thecache control header (it is internally called by thecache_control() decorator):

  1. from django.views.decorators.cache import patch_cache_control
  2. from django.views.decorators.vary import vary_on_cookie
  3.  
  4. @vary_on_cookie
  5. def list_blog_entries_view(request):
  6. if request.user.is_anonymous:
  7. response = render_only_public_entries()
  8. patch_cache_control(response, public=True)
  9. else:
  10. response = render_private_and_public_entries(request.user)
  11. patch_cache_control(response, private=True)
  12.  
  13. return response

You can control downstream caches in other ways as well (see RFC 7234 fordetails on HTTP caching). For example, even if you don’t use Django’sserver-side cache framework, you can still tell clients to cache a view for acertain amount of time with the max-agedirective:

  1. from django.views.decorators.cache import cache_control
  2.  
  3. @cache_control(max_age=3600)
  4. def my_view(request):
  5. ...

(If you do use the caching middleware, it already sets the max-age withthe value of the CACHE_MIDDLEWARE_SECONDS setting. In that case,the custom max_age from thecache_control() decorator will takeprecedence, and the header values will be merged correctly.)

Any valid Cache-Control response directive is valid in cache_control().Here are some more examples:

  • no_transform=True
  • must_revalidate=True
  • stale_while_revalidate=num_secondsThe full list of known directives can be found in the IANA registry(note that not all of them apply to responses).

If you want to use headers to disable caching altogether,never_cache() is a view decorator thatadds headers to ensure the response won’t be cached by browsers or othercaches. Example:

  1. from django.views.decorators.cache import never_cache
  2.  
  3. @never_cache
  4. def myview(request):
  5. ...

Order of MIDDLEWARE

If you use caching middleware, it’s important to put each half in the rightplace within the MIDDLEWARE setting. That’s because the cachemiddleware needs to know which headers by which to vary the cache storage.Middleware always adds something to the Vary response header when it can.

UpdateCacheMiddleware runs during the response phase, where middleware isrun in reverse order, so an item at the top of the list runs last during theresponse phase. Thus, you need to make sure that UpdateCacheMiddlewareappears before any other middleware that might add something to the Varyheader. The following middleware modules do so:

  • SessionMiddleware adds Cookie
  • GZipMiddleware adds Accept-Encoding
  • LocaleMiddleware adds Accept-LanguageFetchFromCacheMiddleware, on the other hand, runs during the request phase,where middleware is applied first-to-last, so an item at the top of the listruns first during the request phase. The FetchFromCacheMiddleware alsoneeds to run after other middleware updates the Vary header, soFetchFromCacheMiddleware must be after any item that does so.