mirror of
https://github.com/jazzband/django-constance.git
synced 2026-04-17 21:50:59 +00:00
197 lines
6.6 KiB
Python
197 lines
6.6 KiB
Python
import asyncio
|
|
from threading import RLock
|
|
from time import monotonic
|
|
|
|
from django.core.exceptions import ImproperlyConfigured
|
|
|
|
from constance import config
|
|
from constance import settings
|
|
from constance import signals
|
|
from constance import utils
|
|
from constance.backends import Backend
|
|
from constance.codecs import dumps
|
|
from constance.codecs import loads
|
|
|
|
|
|
class RedisBackend(Backend):
|
|
def __init__(self):
|
|
super().__init__()
|
|
self._prefix = settings.REDIS_PREFIX
|
|
connection_cls = settings.REDIS_CONNECTION_CLASS
|
|
if connection_cls is not None:
|
|
self._rd = utils.import_module_attr(connection_cls)()
|
|
self._ard = self._rd
|
|
else:
|
|
try:
|
|
import redis
|
|
except ImportError:
|
|
raise ImproperlyConfigured("The Redis backend requires redis-py to be installed.") from None
|
|
|
|
if isinstance(settings.REDIS_CONNECTION, str):
|
|
self._rd = redis.from_url(settings.REDIS_CONNECTION)
|
|
else:
|
|
self._rd = redis.Redis(**settings.REDIS_CONNECTION)
|
|
|
|
try:
|
|
import redis.asyncio as aredis
|
|
|
|
if isinstance(settings.REDIS_CONNECTION, str):
|
|
self._ard = aredis.from_url(settings.REDIS_CONNECTION)
|
|
else:
|
|
self._ard = aredis.Redis(**settings.REDIS_CONNECTION)
|
|
except ImportError:
|
|
self._ard = self._rd
|
|
|
|
def add_prefix(self, key):
|
|
return f"{self._prefix}{key}"
|
|
|
|
def get(self, key):
|
|
value = self._rd.get(self.add_prefix(key))
|
|
if value:
|
|
return loads(value)
|
|
return None
|
|
|
|
async def aget(self, key):
|
|
if hasattr(self._ard, "aget"):
|
|
value = await self._ard.aget(self.add_prefix(key))
|
|
else:
|
|
value = await asyncio.to_thread(self._rd.get, self.add_prefix(key))
|
|
if value:
|
|
return loads(value)
|
|
return None
|
|
|
|
def mget(self, keys):
|
|
if not keys:
|
|
return
|
|
prefixed_keys = [self.add_prefix(key) for key in keys]
|
|
for key, value in zip(keys, self._rd.mget(prefixed_keys)):
|
|
if value:
|
|
yield key, loads(value)
|
|
|
|
async def amget(self, keys):
|
|
if not keys:
|
|
return {}
|
|
prefixed_keys = [self.add_prefix(key) for key in keys]
|
|
if hasattr(self._ard, "amget"):
|
|
values = await self._ard.amget(prefixed_keys)
|
|
else:
|
|
values = await asyncio.to_thread(self._rd.mget, prefixed_keys)
|
|
return {key: loads(value) for key, value in zip(keys, values) if value}
|
|
|
|
def set(self, key, value):
|
|
old_value = self.get(key)
|
|
self._rd.set(self.add_prefix(key), dumps(value))
|
|
signals.config_updated.send(sender=config, key=key, old_value=old_value, new_value=value)
|
|
|
|
async def aset(self, key, value):
|
|
# We need the old value for the signal.
|
|
# Signals are synchronous in Django, but we can't easily change that here.
|
|
old_value = await self.aget(key)
|
|
if hasattr(self._ard, "aset"):
|
|
await self._ard.aset(self.add_prefix(key), dumps(value))
|
|
else:
|
|
await asyncio.to_thread(self._rd.set, self.add_prefix(key), dumps(value))
|
|
signals.config_updated.send(sender=config, key=key, old_value=old_value, new_value=value)
|
|
|
|
|
|
class CachingRedisBackend(RedisBackend):
|
|
_sentinel = object()
|
|
_lock = RLock()
|
|
_async_lock = None # Lazy-initialized asyncio.Lock
|
|
|
|
def __init__(self):
|
|
super().__init__()
|
|
self._timeout = settings.REDIS_CACHE_TIMEOUT
|
|
self._cache = {}
|
|
self._sentinel = object()
|
|
|
|
def _get_async_lock(self):
|
|
# Lazily create the asyncio lock to avoid issues with event loops
|
|
if self._async_lock is None:
|
|
self._async_lock = asyncio.Lock()
|
|
return self._async_lock
|
|
|
|
def _has_expired(self, value):
|
|
return value[0] <= monotonic()
|
|
|
|
def _cache_value(self, key, new_value):
|
|
self._cache[key] = (monotonic() + self._timeout, new_value)
|
|
|
|
def get(self, key):
|
|
value = self._cache.get(key, self._sentinel)
|
|
|
|
if value is self._sentinel or self._has_expired(value):
|
|
with self._lock:
|
|
new_value = super().get(key)
|
|
self._cache_value(key, new_value)
|
|
return new_value
|
|
|
|
return value[1]
|
|
|
|
async def aget(self, key):
|
|
value = self._cache.get(key, self._sentinel)
|
|
|
|
if value is self._sentinel or self._has_expired(value):
|
|
async with self._get_async_lock():
|
|
# Double-check after acquiring lock
|
|
value = self._cache.get(key, self._sentinel)
|
|
if value is self._sentinel or self._has_expired(value):
|
|
new_value = await super().aget(key)
|
|
self._cache_value(key, new_value)
|
|
return new_value
|
|
return value[1]
|
|
|
|
return value[1]
|
|
|
|
def set(self, key, value):
|
|
with self._lock:
|
|
super().set(key, value)
|
|
self._cache_value(key, value)
|
|
|
|
async def aset(self, key, value):
|
|
async with self._get_async_lock():
|
|
await super().aset(key, value)
|
|
self._cache_value(key, value)
|
|
|
|
def mget(self, keys):
|
|
if not keys:
|
|
return
|
|
for key in keys:
|
|
value = self.get(key)
|
|
if value is not None:
|
|
yield key, value
|
|
|
|
async def amget(self, keys):
|
|
if not keys:
|
|
return {}
|
|
|
|
results = {}
|
|
missing_keys = []
|
|
|
|
# First, check the local cache for all keys
|
|
for key in keys:
|
|
value = self._cache.get(key, self._sentinel)
|
|
if value is not self._sentinel and not self._has_expired(value):
|
|
results[key] = value[1]
|
|
else:
|
|
missing_keys.append(key)
|
|
|
|
# Fetch missing keys from Redis
|
|
if missing_keys:
|
|
async with self._get_async_lock():
|
|
# Re-check cache for keys that might have been fetched while waiting for lock
|
|
still_missing = []
|
|
for key in missing_keys:
|
|
value = self._cache.get(key, self._sentinel)
|
|
if value is not self._sentinel and not self._has_expired(value):
|
|
results[key] = value[1]
|
|
else:
|
|
still_missing.append(key)
|
|
|
|
if still_missing:
|
|
fetched = await super().amget(still_missing)
|
|
for key, value in fetched.items():
|
|
self._cache_value(key, value)
|
|
results[key] = value
|
|
|
|
return results
|