nitter/src/redis_cache.nim

142 lines
4.5 KiB
Nim
Raw Normal View History

import asyncdispatch, times, strutils, tables, hashes
import redis, redpool, flatty, supersnappy
2020-06-01 02:16:24 +02:00
import types, api
const redisNil = "\0\0"
var
2020-06-22 03:50:16 +02:00
pool {.threadvar.}: RedisPool
2020-06-01 02:16:24 +02:00
baseCacheTime = 60 * 60
rssCacheTime: int
listCacheTime*: int
# flatty can't serialize DateTime, so we need to define this
proc toFlatty*(s: var string, x: DateTime) =
s.toFlatty(x.toTime().toUnix())
proc fromFlatty*(s: string, i: var int, x: var DateTime) =
x = fromUnix(s.fromFlatty(int64)).utc()
2020-06-01 02:16:24 +02:00
proc setCacheTimes*(cfg: Config) =
rssCacheTime = cfg.rssCacheTime * 60
listCacheTime = cfg.listCacheTime * 60
proc migrate*(key, match: string) {.async.} =
pool.withAcquire(r):
let hasKey = await r.get(key)
if hasKey == redisNil:
let list = await r.scan(newCursor(0), match, 100000)
r.startPipelining()
for item in list:
2021-12-26 22:59:27 +01:00
discard await r.del(item)
await r.setk(key, "true")
discard await r.flushPipeline()
proc initRedisPool*(cfg: Config) {.async.} =
2020-06-01 13:36:20 +02:00
try:
pool = await newRedisPool(cfg.redisConns, cfg.redisMaxConns,
host=cfg.redisHost, port=cfg.redisPort,
password=cfg.redisPassword)
await migrate("flatty", "*:*")
await migrate("snappyRss", "rss:*")
2021-12-26 22:59:27 +01:00
await migrate("userBuckets", "p:*")
await migrate("profileDates", "p:*")
pool.withAcquire(r):
2021-12-26 22:59:27 +01:00
# optimize memory usage for profile ID buckets
await r.configSet("hash-max-ziplist-entries", "1000")
2020-06-01 13:36:20 +02:00
except OSError:
2020-06-07 09:18:40 +02:00
stdout.write "Failed to connect to Redis.\n"
stdout.flushFile
quit(1)
2020-06-01 02:16:24 +02:00
template pidKey(name: string): string = "pid:" & $(hash(name) div 1_000_000)
template profileKey(name: string): string = "p:" & name
template listKey(l: List): string = toLower("l:" & l.username & '/' & l.name)
2020-06-01 02:16:24 +02:00
proc get(query: string): Future[string] {.async.} =
pool.withAcquire(r):
result = await r.get(query)
2020-06-02 22:36:02 +02:00
proc setex(key: string; time: int; data: string) {.async.} =
2020-06-01 02:16:24 +02:00
pool.withAcquire(r):
2020-06-02 22:36:02 +02:00
discard await r.setex(key, time, data)
2020-06-01 02:16:24 +02:00
2020-06-02 22:36:02 +02:00
proc cache*(data: List) {.async.} =
await setex(data.listKey, listCacheTime, compress(toFlatty(data)))
2020-06-01 02:16:24 +02:00
2020-06-17 00:20:34 +02:00
proc cache*(data: PhotoRail; name: string) {.async.} =
await setex("pr:" & name, baseCacheTime, compress(toFlatty(data)))
2020-06-01 02:16:24 +02:00
2020-06-02 22:36:02 +02:00
proc cache*(data: Profile) {.async.} =
if data.username.len == 0 or data.id.len == 0: return
let name = toLower(data.username)
2020-06-01 02:16:24 +02:00
pool.withAcquire(r):
r.startPipelining()
discard await r.setex(name.profileKey, baseCacheTime, compress(toFlatty(data)))
discard await r.hset(name.pidKey, name, data.id)
2020-06-01 02:16:24 +02:00
discard await r.flushPipeline()
2020-06-03 00:03:41 +02:00
proc cacheProfileId*(username, id: string) {.async.} =
if username.len == 0 or id.len == 0: return
let name = toLower(username)
2020-06-03 00:03:41 +02:00
pool.withAcquire(r):
discard await r.hset(name.pidKey, name, id)
2020-06-03 00:03:41 +02:00
proc cacheRss*(query: string; rss: Rss) {.async.} =
2020-06-01 02:16:24 +02:00
let key = "rss:" & query
pool.withAcquire(r):
r.startPipelining()
discard await r.hset(key, "rss", rss.feed)
discard await r.hset(key, "min", rss.cursor)
2020-06-01 02:16:24 +02:00
discard await r.expire(key, rssCacheTime)
discard await r.flushPipeline()
proc getProfileId*(username: string): Future[string] {.async.} =
let name = toLower(username)
2020-06-01 02:16:24 +02:00
pool.withAcquire(r):
result = await r.hget(name.pidKey, name)
2020-06-01 02:16:24 +02:00
if result == redisNil:
result.setLen(0)
proc getCachedProfile*(username: string; fetch=true): Future[Profile] {.async.} =
2020-06-01 02:16:24 +02:00
let prof = await get("p:" & toLower(username))
if prof != redisNil:
result = fromFlatty(uncompress(prof), Profile)
2020-06-03 00:03:41 +02:00
elif fetch:
result = await getProfile(username)
2020-06-01 02:16:24 +02:00
2020-06-17 00:20:34 +02:00
proc getCachedPhotoRail*(name: string): Future[PhotoRail] {.async.} =
if name.len == 0: return
let rail = await get("pr:" & toLower(name))
2020-06-01 02:16:24 +02:00
if rail != redisNil:
result = fromFlatty(uncompress(rail), PhotoRail)
2020-06-01 02:16:24 +02:00
else:
2020-06-17 00:20:34 +02:00
result = await getPhotoRail(name)
await cache(result, name)
2020-06-01 02:16:24 +02:00
proc getCachedList*(username=""; name=""; id=""): Future[List] {.async.} =
let list = if id.len > 0: redisNil
else: await get(toLower("l:" & username & '/' & name))
if list != redisNil:
result = fromFlatty(uncompress(list), List)
2020-06-01 02:16:24 +02:00
else:
if id.len > 0:
result = await getGraphListById(id)
else:
result = await getGraphList(username, name)
2020-06-02 22:36:02 +02:00
await cache(result)
2020-06-01 02:16:24 +02:00
proc getCachedRss*(key: string): Future[Rss] {.async.} =
let k = "rss:" & key
2020-06-01 02:16:24 +02:00
pool.withAcquire(r):
result.cursor = await r.hget(k, "min")
if result.cursor.len > 2:
result.feed = await r.hget(k, "rss")
else:
result.cursor.setLen 0