nfsd: remove recurring workqueue job to clean DRC

We have a shrinker, we clean out the cache when nfsd is shut down, and
prune the chains on each request. A recurring workqueue job seems like
unnecessary overhead. Just remove it.

Signed-off-by: Jeff Layton <jeff.layton@primarydata.com>
Signed-off-by: J. Bruce Fields <bfields@redhat.com>
This commit is contained in:
Jeff Layton 2015-11-04 11:02:29 -05:00 committed by J. Bruce Fields
parent ea833f5de3
commit 3e80dbcda7
1 changed files with 0 additions and 26 deletions

View File

@ -63,7 +63,6 @@ static unsigned int longest_chain;
static unsigned int longest_chain_cachesize; static unsigned int longest_chain_cachesize;
static int nfsd_cache_append(struct svc_rqst *rqstp, struct kvec *vec); static int nfsd_cache_append(struct svc_rqst *rqstp, struct kvec *vec);
static void cache_cleaner_func(struct work_struct *unused);
static unsigned long nfsd_reply_cache_count(struct shrinker *shrink, static unsigned long nfsd_reply_cache_count(struct shrinker *shrink,
struct shrink_control *sc); struct shrink_control *sc);
static unsigned long nfsd_reply_cache_scan(struct shrinker *shrink, static unsigned long nfsd_reply_cache_scan(struct shrinker *shrink,
@ -75,13 +74,6 @@ static struct shrinker nfsd_reply_cache_shrinker = {
.seeks = 1, .seeks = 1,
}; };
/*
* locking for the reply cache:
* A cache entry is "single use" if c_state == RC_INPROG
* Otherwise, it when accessing _prev or _next, the lock must be held.
*/
static DECLARE_DELAYED_WORK(cache_cleaner, cache_cleaner_func);
/* /*
* Put a cap on the size of the DRC based on the amount of available * Put a cap on the size of the DRC based on the amount of available
* low memory in the machine. * low memory in the machine.
@ -203,7 +195,6 @@ void nfsd_reply_cache_shutdown(void)
unsigned int i; unsigned int i;
unregister_shrinker(&nfsd_reply_cache_shrinker); unregister_shrinker(&nfsd_reply_cache_shrinker);
cancel_delayed_work_sync(&cache_cleaner);
for (i = 0; i < drc_hashsize; i++) { for (i = 0; i < drc_hashsize; i++) {
struct list_head *head = &drc_hashtbl[i].lru_head; struct list_head *head = &drc_hashtbl[i].lru_head;
@ -230,7 +221,6 @@ lru_put_end(struct nfsd_drc_bucket *b, struct svc_cacherep *rp)
{ {
rp->c_timestamp = jiffies; rp->c_timestamp = jiffies;
list_move_tail(&rp->c_lru, &b->lru_head); list_move_tail(&rp->c_lru, &b->lru_head);
schedule_delayed_work(&cache_cleaner, RC_EXPIRE);
} }
static long static long
@ -264,7 +254,6 @@ prune_cache_entries(void)
{ {
unsigned int i; unsigned int i;
long freed = 0; long freed = 0;
bool cancel = true;
for (i = 0; i < drc_hashsize; i++) { for (i = 0; i < drc_hashsize; i++) {
struct nfsd_drc_bucket *b = &drc_hashtbl[i]; struct nfsd_drc_bucket *b = &drc_hashtbl[i];
@ -273,26 +262,11 @@ prune_cache_entries(void)
continue; continue;
spin_lock(&b->cache_lock); spin_lock(&b->cache_lock);
freed += prune_bucket(b); freed += prune_bucket(b);
if (!list_empty(&b->lru_head))
cancel = false;
spin_unlock(&b->cache_lock); spin_unlock(&b->cache_lock);
} }
/*
* Conditionally rearm the job to run in RC_EXPIRE since we just
* ran the pruner.
*/
if (!cancel)
mod_delayed_work(system_wq, &cache_cleaner, RC_EXPIRE);
return freed; return freed;
} }
static void
cache_cleaner_func(struct work_struct *unused)
{
prune_cache_entries();
}
static unsigned long static unsigned long
nfsd_reply_cache_count(struct shrinker *shrink, struct shrink_control *sc) nfsd_reply_cache_count(struct shrinker *shrink, struct shrink_control *sc)
{ {