Commit 375d80fe authored by Kazuhiko Shiozaki's avatar Kazuhiko Shiozaki

* differenciate MemcachedDict for different ERP5 instances by using SharedDict...

* differenciate MemcachedDict for different ERP5 instances by using SharedDict instead of MemcachedDict just same as portal_memcached.getMemcachedDict().
* since encoding key is done in MemcachedTool level, there is no need to do the similar thing in DistributedRamCache level.
* stop diffentiate caches by cache plugin url, because it makes the key too long and it still cannot prevent cache duplication when a user modifies the configuration of cache plugins.


git-svn-id: https://svn.erp5.org/repos/public/erp5/trunk@29534 20353a03-c40f-0410-a6d1-a30d3c3de9de
parent 08c55c3f
...@@ -40,7 +40,7 @@ from base64 import encodestring ...@@ -40,7 +40,7 @@ from base64 import encodestring
try: try:
import memcache import memcache
from Products.ERP5Type.Tool.MemcachedTool import MemcachedDict from Products.ERP5Type.Tool.MemcachedTool import MemcachedDict, SharedDict
except ImportError: except ImportError:
LOG('DistributedRamCache', 0, 'unable to import memcache') LOG('DistributedRamCache', 0, 'unable to import memcache')
...@@ -62,7 +62,6 @@ class DistributedRamCache(BaseCache): ...@@ -62,7 +62,6 @@ class DistributedRamCache(BaseCache):
self._server_max_value_length = params.get('server_max_value_length', 1024*1024) self._server_max_value_length = params.get('server_max_value_length', 1024*1024)
self._debug_level = params.get('debug_level', 0) self._debug_level = params.get('debug_level', 0)
self._key_prefix = params.get('key_prefix', '') self._key_prefix = params.get('key_prefix', '')
self._cache_plugin_path = params.get('cache_plugin_path')
BaseCache.__init__(self) BaseCache.__init__(self)
def initCacheStorage(self): def initCacheStorage(self):
...@@ -78,26 +77,20 @@ class DistributedRamCache(BaseCache): ...@@ -78,26 +77,20 @@ class DistributedRamCache(BaseCache):
try: try:
dictionary = connection_pool.memcached_dict dictionary = connection_pool.memcached_dict
except AttributeError: except AttributeError:
dictionary = MemcachedDict(self._servers.split('\n'), dictionary = SharedDict(
server_max_key_length=self._server_max_key_length, MemcachedDict(self._servers.split('\n'),
server_max_value_length=self._server_max_value_length) server_max_key_length=self._server_max_key_length,
server_max_value_length=self._server_max_value_length),
prefix=self._key_prefix)
connection_pool.memcached_dict = dictionary connection_pool.memcached_dict = dictionary
return dictionary return dictionary
def checkAndFixCacheId(self, cache_id, scope): def _getCacheId(self, cache_id, scope):
## memcached doesn't support namespaces (cache scopes) so to "emmulate" return '%s_%s' % (scope, cache_id)
## such behaviour when constructing cache_id we add scope in front
cache_id = "%s%s%s%s" % (self._key_prefix, self._cache_plugin_path,
scope, cache_id)
# Escape key to normalise some chars
# which are not allowed by memcached
# Could reach the limit of max_key_len
cache_id = encodestring(cache_id).replace('\n', '')
return cache_id
def get(self, cache_id, scope, default=_MARKER): def get(self, cache_id, scope, default=_MARKER):
cache_storage = self.getCacheStorage() cache_storage = self.getCacheStorage()
cache_id = self.checkAndFixCacheId(cache_id, scope) cache_id = self._getCacheId(cache_id, scope)
cache_entry = cache_storage.get(cache_id) cache_entry = cache_storage.get(cache_id)
#Simulate the behaviour of a standard Dictionary #Simulate the behaviour of a standard Dictionary
if not isinstance(cache_entry, CacheEntry): if not isinstance(cache_entry, CacheEntry):
...@@ -116,7 +109,7 @@ class DistributedRamCache(BaseCache): ...@@ -116,7 +109,7 @@ class DistributedRamCache(BaseCache):
def set(self, cache_id, scope, value, cache_duration=None, calculation_time=0): def set(self, cache_id, scope, value, cache_duration=None, calculation_time=0):
cache_storage = self.getCacheStorage() cache_storage = self.getCacheStorage()
cache_id = self.checkAndFixCacheId(cache_id, scope) cache_id = self._getCacheId(cache_id, scope)
cache_entry = CacheEntry(value, cache_duration, calculation_time) cache_entry = CacheEntry(value, cache_duration, calculation_time)
cache_storage.set(cache_id, cache_entry) cache_storage.set(cache_id, cache_entry)
self.markCacheMiss() self.markCacheMiss()
...@@ -134,12 +127,12 @@ class DistributedRamCache(BaseCache): ...@@ -134,12 +127,12 @@ class DistributedRamCache(BaseCache):
def delete(self, cache_id, scope): def delete(self, cache_id, scope):
cache_storage = self.getCacheStorage() cache_storage = self.getCacheStorage()
cache_id = self.checkAndFixCacheId(cache_id, scope) cache_id = self._getCacheId(cache_id, scope)
del cache_storage[cache_id] del cache_storage[cache_id]
def has_key(self, cache_id, scope): def has_key(self, cache_id, scope):
cache_storage = self.getCacheStorage() cache_storage = self.getCacheStorage()
cache_id = self.checkAndFixCacheId(cache_id, scope) cache_id = self._getCacheId(cache_id, scope)
cache_entry = cache_storage.get(cache_id) cache_entry = cache_storage.get(cache_id)
to_return = False to_return = False
if isinstance(cache_entry, CacheEntry): if isinstance(cache_entry, CacheEntry):
......
...@@ -91,7 +91,6 @@ class CacheTool(BaseTool): ...@@ -91,7 +91,6 @@ class CacheTool(BaseTool):
'server': server, 'server': server,
'server_max_key_length': memcached_plugin.getServerMaxKeyLength(), 'server_max_key_length': memcached_plugin.getServerMaxKeyLength(),
'server_max_value_length': memcached_plugin.getServerMaxValueLength(), 'server_max_value_length': memcached_plugin.getServerMaxValueLength(),
'cache_plugin_path': cp.getPath(),
'key_prefix': getattr(self, 'erp5_site_global_id', '') 'key_prefix': getattr(self, 'erp5_site_global_id', '')
} }
cache_obj = DistributedRamCache(init_dict) cache_obj = DistributedRamCache(init_dict)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment