[PATCH] Manage jbd allocations from its own slabs
JBD currently allocates commit and frozen buffers from slabs. With CONFIG_SLAB_DEBUG, its possible for an allocation to cross the page boundary causing IO problems. https://bugzilla.redhat.com/bugzilla/show_bug.cgi?id=200127 So, instead of allocating these from regular slabs - manage allocation from its own slabs and disable slab debug for these slabs. [akpm@osdl.org: cleanups] Signed-off-by: Badari Pulavarty <pbadari@us.ibm.com> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
parent
4c4d50f7b3
commit
ea817398e6
|
@ -261,7 +261,7 @@ void journal_commit_transaction(journal_t *journal)
|
||||||
struct buffer_head *bh = jh2bh(jh);
|
struct buffer_head *bh = jh2bh(jh);
|
||||||
|
|
||||||
jbd_lock_bh_state(bh);
|
jbd_lock_bh_state(bh);
|
||||||
kfree(jh->b_committed_data);
|
jbd_slab_free(jh->b_committed_data, bh->b_size);
|
||||||
jh->b_committed_data = NULL;
|
jh->b_committed_data = NULL;
|
||||||
jbd_unlock_bh_state(bh);
|
jbd_unlock_bh_state(bh);
|
||||||
}
|
}
|
||||||
|
@ -745,14 +745,14 @@ restart_loop:
|
||||||
* Otherwise, we can just throw away the frozen data now.
|
* Otherwise, we can just throw away the frozen data now.
|
||||||
*/
|
*/
|
||||||
if (jh->b_committed_data) {
|
if (jh->b_committed_data) {
|
||||||
kfree(jh->b_committed_data);
|
jbd_slab_free(jh->b_committed_data, bh->b_size);
|
||||||
jh->b_committed_data = NULL;
|
jh->b_committed_data = NULL;
|
||||||
if (jh->b_frozen_data) {
|
if (jh->b_frozen_data) {
|
||||||
jh->b_committed_data = jh->b_frozen_data;
|
jh->b_committed_data = jh->b_frozen_data;
|
||||||
jh->b_frozen_data = NULL;
|
jh->b_frozen_data = NULL;
|
||||||
}
|
}
|
||||||
} else if (jh->b_frozen_data) {
|
} else if (jh->b_frozen_data) {
|
||||||
kfree(jh->b_frozen_data);
|
jbd_slab_free(jh->b_frozen_data, bh->b_size);
|
||||||
jh->b_frozen_data = NULL;
|
jh->b_frozen_data = NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -84,6 +84,7 @@ EXPORT_SYMBOL(journal_force_commit);
|
||||||
|
|
||||||
static int journal_convert_superblock_v1(journal_t *, journal_superblock_t *);
|
static int journal_convert_superblock_v1(journal_t *, journal_superblock_t *);
|
||||||
static void __journal_abort_soft (journal_t *journal, int errno);
|
static void __journal_abort_soft (journal_t *journal, int errno);
|
||||||
|
static int journal_create_jbd_slab(size_t slab_size);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Helper function used to manage commit timeouts
|
* Helper function used to manage commit timeouts
|
||||||
|
@ -328,10 +329,10 @@ repeat:
|
||||||
char *tmp;
|
char *tmp;
|
||||||
|
|
||||||
jbd_unlock_bh_state(bh_in);
|
jbd_unlock_bh_state(bh_in);
|
||||||
tmp = jbd_rep_kmalloc(bh_in->b_size, GFP_NOFS);
|
tmp = jbd_slab_alloc(bh_in->b_size, GFP_NOFS);
|
||||||
jbd_lock_bh_state(bh_in);
|
jbd_lock_bh_state(bh_in);
|
||||||
if (jh_in->b_frozen_data) {
|
if (jh_in->b_frozen_data) {
|
||||||
kfree(tmp);
|
jbd_slab_free(tmp, bh_in->b_size);
|
||||||
goto repeat;
|
goto repeat;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1069,17 +1070,17 @@ static int load_superblock(journal_t *journal)
|
||||||
int journal_load(journal_t *journal)
|
int journal_load(journal_t *journal)
|
||||||
{
|
{
|
||||||
int err;
|
int err;
|
||||||
|
journal_superblock_t *sb;
|
||||||
|
|
||||||
err = load_superblock(journal);
|
err = load_superblock(journal);
|
||||||
if (err)
|
if (err)
|
||||||
return err;
|
return err;
|
||||||
|
|
||||||
|
sb = journal->j_superblock;
|
||||||
/* If this is a V2 superblock, then we have to check the
|
/* If this is a V2 superblock, then we have to check the
|
||||||
* features flags on it. */
|
* features flags on it. */
|
||||||
|
|
||||||
if (journal->j_format_version >= 2) {
|
if (journal->j_format_version >= 2) {
|
||||||
journal_superblock_t *sb = journal->j_superblock;
|
|
||||||
|
|
||||||
if ((sb->s_feature_ro_compat &
|
if ((sb->s_feature_ro_compat &
|
||||||
~cpu_to_be32(JFS_KNOWN_ROCOMPAT_FEATURES)) ||
|
~cpu_to_be32(JFS_KNOWN_ROCOMPAT_FEATURES)) ||
|
||||||
(sb->s_feature_incompat &
|
(sb->s_feature_incompat &
|
||||||
|
@ -1090,6 +1091,13 @@ int journal_load(journal_t *journal)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Create a slab for this blocksize
|
||||||
|
*/
|
||||||
|
err = journal_create_jbd_slab(cpu_to_be32(sb->s_blocksize));
|
||||||
|
if (err)
|
||||||
|
return err;
|
||||||
|
|
||||||
/* Let the recovery code check whether it needs to recover any
|
/* Let the recovery code check whether it needs to recover any
|
||||||
* data from the journal. */
|
* data from the journal. */
|
||||||
if (journal_recover(journal))
|
if (journal_recover(journal))
|
||||||
|
@ -1611,6 +1619,77 @@ void * __jbd_kmalloc (const char *where, size_t size, gfp_t flags, int retry)
|
||||||
return kmalloc(size, flags | (retry ? __GFP_NOFAIL : 0));
|
return kmalloc(size, flags | (retry ? __GFP_NOFAIL : 0));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* jbd slab management: create 1k, 2k, 4k, 8k slabs as needed
|
||||||
|
* and allocate frozen and commit buffers from these slabs.
|
||||||
|
*
|
||||||
|
* Reason for doing this is to avoid, SLAB_DEBUG - since it could
|
||||||
|
* cause bh to cross page boundary.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#define JBD_MAX_SLABS 5
|
||||||
|
#define JBD_SLAB_INDEX(size) (size >> 11)
|
||||||
|
|
||||||
|
static kmem_cache_t *jbd_slab[JBD_MAX_SLABS];
|
||||||
|
static const char *jbd_slab_names[JBD_MAX_SLABS] = {
|
||||||
|
"jbd_1k", "jbd_2k", "jbd_4k", NULL, "jbd_8k"
|
||||||
|
};
|
||||||
|
|
||||||
|
static void journal_destroy_jbd_slabs(void)
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
|
||||||
|
for (i = 0; i < JBD_MAX_SLABS; i++) {
|
||||||
|
if (jbd_slab[i])
|
||||||
|
kmem_cache_destroy(jbd_slab[i]);
|
||||||
|
jbd_slab[i] = NULL;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static int journal_create_jbd_slab(size_t slab_size)
|
||||||
|
{
|
||||||
|
int i = JBD_SLAB_INDEX(slab_size);
|
||||||
|
|
||||||
|
BUG_ON(i >= JBD_MAX_SLABS);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Check if we already have a slab created for this size
|
||||||
|
*/
|
||||||
|
if (jbd_slab[i])
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Create a slab and force alignment to be same as slabsize -
|
||||||
|
* this will make sure that allocations won't cross the page
|
||||||
|
* boundary.
|
||||||
|
*/
|
||||||
|
jbd_slab[i] = kmem_cache_create(jbd_slab_names[i],
|
||||||
|
slab_size, slab_size, 0, NULL, NULL);
|
||||||
|
if (!jbd_slab[i]) {
|
||||||
|
printk(KERN_EMERG "JBD: no memory for jbd_slab cache\n");
|
||||||
|
return -ENOMEM;
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
void * jbd_slab_alloc(size_t size, gfp_t flags)
|
||||||
|
{
|
||||||
|
int idx;
|
||||||
|
|
||||||
|
idx = JBD_SLAB_INDEX(size);
|
||||||
|
BUG_ON(jbd_slab[idx] == NULL);
|
||||||
|
return kmem_cache_alloc(jbd_slab[idx], flags | __GFP_NOFAIL);
|
||||||
|
}
|
||||||
|
|
||||||
|
void jbd_slab_free(void *ptr, size_t size)
|
||||||
|
{
|
||||||
|
int idx;
|
||||||
|
|
||||||
|
idx = JBD_SLAB_INDEX(size);
|
||||||
|
BUG_ON(jbd_slab[idx] == NULL);
|
||||||
|
kmem_cache_free(jbd_slab[idx], ptr);
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Journal_head storage management
|
* Journal_head storage management
|
||||||
*/
|
*/
|
||||||
|
@ -1799,13 +1878,13 @@ static void __journal_remove_journal_head(struct buffer_head *bh)
|
||||||
printk(KERN_WARNING "%s: freeing "
|
printk(KERN_WARNING "%s: freeing "
|
||||||
"b_frozen_data\n",
|
"b_frozen_data\n",
|
||||||
__FUNCTION__);
|
__FUNCTION__);
|
||||||
kfree(jh->b_frozen_data);
|
jbd_slab_free(jh->b_frozen_data, bh->b_size);
|
||||||
}
|
}
|
||||||
if (jh->b_committed_data) {
|
if (jh->b_committed_data) {
|
||||||
printk(KERN_WARNING "%s: freeing "
|
printk(KERN_WARNING "%s: freeing "
|
||||||
"b_committed_data\n",
|
"b_committed_data\n",
|
||||||
__FUNCTION__);
|
__FUNCTION__);
|
||||||
kfree(jh->b_committed_data);
|
jbd_slab_free(jh->b_committed_data, bh->b_size);
|
||||||
}
|
}
|
||||||
bh->b_private = NULL;
|
bh->b_private = NULL;
|
||||||
jh->b_bh = NULL; /* debug, really */
|
jh->b_bh = NULL; /* debug, really */
|
||||||
|
@ -1961,6 +2040,7 @@ static void journal_destroy_caches(void)
|
||||||
journal_destroy_revoke_caches();
|
journal_destroy_revoke_caches();
|
||||||
journal_destroy_journal_head_cache();
|
journal_destroy_journal_head_cache();
|
||||||
journal_destroy_handle_cache();
|
journal_destroy_handle_cache();
|
||||||
|
journal_destroy_jbd_slabs();
|
||||||
}
|
}
|
||||||
|
|
||||||
static int __init journal_init(void)
|
static int __init journal_init(void)
|
||||||
|
|
|
@ -666,8 +666,9 @@ repeat:
|
||||||
if (!frozen_buffer) {
|
if (!frozen_buffer) {
|
||||||
JBUFFER_TRACE(jh, "allocate memory for buffer");
|
JBUFFER_TRACE(jh, "allocate memory for buffer");
|
||||||
jbd_unlock_bh_state(bh);
|
jbd_unlock_bh_state(bh);
|
||||||
frozen_buffer = jbd_kmalloc(jh2bh(jh)->b_size,
|
frozen_buffer =
|
||||||
GFP_NOFS);
|
jbd_slab_alloc(jh2bh(jh)->b_size,
|
||||||
|
GFP_NOFS);
|
||||||
if (!frozen_buffer) {
|
if (!frozen_buffer) {
|
||||||
printk(KERN_EMERG
|
printk(KERN_EMERG
|
||||||
"%s: OOM for frozen_buffer\n",
|
"%s: OOM for frozen_buffer\n",
|
||||||
|
@ -879,7 +880,7 @@ int journal_get_undo_access(handle_t *handle, struct buffer_head *bh)
|
||||||
|
|
||||||
repeat:
|
repeat:
|
||||||
if (!jh->b_committed_data) {
|
if (!jh->b_committed_data) {
|
||||||
committed_data = jbd_kmalloc(jh2bh(jh)->b_size, GFP_NOFS);
|
committed_data = jbd_slab_alloc(jh2bh(jh)->b_size, GFP_NOFS);
|
||||||
if (!committed_data) {
|
if (!committed_data) {
|
||||||
printk(KERN_EMERG "%s: No memory for committed data\n",
|
printk(KERN_EMERG "%s: No memory for committed data\n",
|
||||||
__FUNCTION__);
|
__FUNCTION__);
|
||||||
|
@ -906,7 +907,7 @@ repeat:
|
||||||
out:
|
out:
|
||||||
journal_put_journal_head(jh);
|
journal_put_journal_head(jh);
|
||||||
if (unlikely(committed_data))
|
if (unlikely(committed_data))
|
||||||
kfree(committed_data);
|
jbd_slab_free(committed_data, bh->b_size);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -72,6 +72,9 @@ extern int journal_enable_debug;
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
extern void * __jbd_kmalloc (const char *where, size_t size, gfp_t flags, int retry);
|
extern void * __jbd_kmalloc (const char *where, size_t size, gfp_t flags, int retry);
|
||||||
|
extern void * jbd_slab_alloc(size_t size, gfp_t flags);
|
||||||
|
extern void jbd_slab_free(void *ptr, size_t size);
|
||||||
|
|
||||||
#define jbd_kmalloc(size, flags) \
|
#define jbd_kmalloc(size, flags) \
|
||||||
__jbd_kmalloc(__FUNCTION__, (size), (flags), journal_oom_retry)
|
__jbd_kmalloc(__FUNCTION__, (size), (flags), journal_oom_retry)
|
||||||
#define jbd_rep_kmalloc(size, flags) \
|
#define jbd_rep_kmalloc(size, flags) \
|
||||||
|
|
Loading…
Reference in New Issue