2
0
mirror of https://github.com/koreader/koreader synced 2024-11-20 03:25:34 +00:00
koreader/frontend/cache.lua

194 lines
5.6 KiB
Lua
Raw Normal View History

2012-05-18 22:35:09 +00:00
--[[
A global LRU cache
]]--
2017-04-29 14:30:16 +00:00
local DataStorage = require("datastorage")
2017-04-29 14:30:16 +00:00
local lfs = require("libs/libkoreader-lfs")
local logger = require("logger")
2017-04-29 14:30:16 +00:00
local md5 = require("ffi/MD5")
if require("device"):isAndroid() then
require("jit").off(true, true)
end
local function calcFreeMem()
2014-03-13 13:52:43 +00:00
local meminfo = io.open("/proc/meminfo", "r")
local freemem = 0
if meminfo then
for line in meminfo:lines() do
local free, buffer, cached, n
free, n = line:gsub("^MemFree:%s-(%d+) kB", "%1")
if n ~= 0 then freemem = freemem + tonumber(free)*1024 end
buffer, n = line:gsub("^Buffers:%s-(%d+) kB", "%1")
if n ~= 0 then freemem = freemem + tonumber(buffer)*1024 end
cached, n = line:gsub("^Cached:%s-(%d+) kB", "%1")
if n ~= 0 then freemem = freemem + tonumber(cached)*1024 end
end
meminfo:close()
end
return freemem
end
local function calcCacheMemSize()
2014-03-13 13:52:43 +00:00
local min = DGLOBAL_CACHE_SIZE_MINIMUM
local max = DGLOBAL_CACHE_SIZE_MAXIMUM
local calc = calcFreeMem()*(DGLOBAL_CACHE_FREE_PROPORTION or 0)
return math.min(max, math.max(min, calc))
end
local cache_path = DataStorage:getDataDir() .. "/cache/"
--[[
-- return a snapshot of disk cached items for subsequent check
--]]
local function getDiskCache()
local cached = {}
for key_md5 in lfs.dir(cache_path) do
local file = cache_path..key_md5
if lfs.attributes(file, "mode") == "file" then
cached[key_md5] = file
end
end
return cached
end
2013-10-18 20:38:07 +00:00
local Cache = {
2014-03-13 13:52:43 +00:00
-- cache configuration:
max_memsize = calcCacheMemSize(),
-- cache state:
current_memsize = 0,
-- associative cache
cache = {},
-- this will hold the LRU order of the cache
cache_order = {},
-- disk Cache snapshot
cached = getDiskCache(),
2012-05-18 22:35:09 +00:00
}
function Cache:new(o)
2014-03-13 13:52:43 +00:00
o = o or {}
setmetatable(o, self)
self.__index = self
return o
end
-- internal: remove reference in cache_order list
function Cache:_unref(key)
for i = #self.cache_order, 1, -1 do
if self.cache_order[i] == key then
table.remove(self.cache_order, i)
end
end
end
-- internal: free cache item
function Cache:_free(key)
if not self.cache[key] then return end
self.current_memsize = self.current_memsize - self.cache[key].size
self.cache[key]:onFree()
self.cache[key] = nil
end
-- drop an item named via key from the cache
function Cache:drop(key)
self:_unref(key)
self:_free(key)
end
2012-05-18 22:35:09 +00:00
function Cache:insert(key, object)
-- make sure that one key only exists once: delete existing
self:drop(key)
2014-03-13 13:52:43 +00:00
-- guarantee that we have enough memory in cache
2016-04-19 06:50:36 +00:00
if (object.size > self.max_memsize) then
logger.warn("too much memory claimed for", key)
2014-10-23 05:48:45 +00:00
return
2014-03-13 13:52:43 +00:00
end
-- delete objects that least recently used
-- (they are at the end of the cache_order array)
while self.current_memsize + object.size > self.max_memsize do
local removed_key = table.remove(self.cache_order)
self:_free(removed_key)
2014-03-13 13:52:43 +00:00
end
-- insert new object in front of the LRU order
table.insert(self.cache_order, 1, key)
self.cache[key] = object
self.current_memsize = self.current_memsize + object.size
2012-05-18 22:35:09 +00:00
end
--[[
-- check for cache item for key
-- if ItemClass is given, disk cache is also checked.
--]]
function Cache:check(key, ItemClass)
2014-03-13 13:52:43 +00:00
if self.cache[key] then
if self.cache_order[1] ~= key then
-- put key in front of the LRU list
self:_unref(key)
2014-03-13 13:52:43 +00:00
table.insert(self.cache_order, 1, key)
end
return self.cache[key]
elseif ItemClass then
local cached = self.cached[md5.sum(key)]
if cached then
local item = ItemClass:new{}
local ok, msg = pcall(item.load, item, cached)
if ok then
self:insert(key, item)
return item
else
logger.warn("discard cache", msg)
end
end
2014-03-13 13:52:43 +00:00
end
2012-05-18 22:35:09 +00:00
end
function Cache:willAccept(size)
2014-03-13 13:52:43 +00:00
-- we only allow single objects to fill 75% of the cache
if size*4 < self.max_memsize*3 then
return true
end
2012-05-18 22:35:09 +00:00
end
function Cache:serialize()
-- calculate disk cache size
local cached_size = 0
local sorted_caches = {}
for _,file in pairs(self.cached) do
table.insert(sorted_caches, {file=file, time=lfs.attributes(file, "access")})
cached_size = cached_size + (lfs.attributes(file, "size") or 0)
end
table.sort(sorted_caches, function(v1,v2) return v1.time > v2.time end)
-- only serialize the most recently used cache
local cache_size = 0
for _, key in ipairs(self.cache_order) do
local cache_item = self.cache[key]
-- only dump cache item that requests serialization explicitly
if cache_item.persistent and cache_item.dump then
logger.dbg("dump cache item", key)
cache_size = cache_item:dump(cache_path..md5.sum(key)) or 0
if cache_size > 0 then break end
end
end
-- set disk cache the same limit as memory cache
while cached_size + cache_size - self.max_memsize > 0 do
-- discard the least recently used cache
local discarded = table.remove(sorted_caches)
cached_size = cached_size - lfs.attributes(discarded.file, "size")
os.remove(discarded.file)
end
-- disk cache may have changes so need to refresh disk cache snapshot
self.cached = getDiskCache()
end
2012-05-18 22:35:09 +00:00
-- blank the cache
function Cache:clear()
2014-03-13 13:52:43 +00:00
for k, _ in pairs(self.cache) do
self.cache[k]:onFree()
end
self.cache = {}
self.cache_order = {}
self.current_memsize = 0
2012-05-18 22:35:09 +00:00
end
2013-10-18 20:38:07 +00:00
return Cache