2019-05-19 14:08:55 +02:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-only
|
2016-12-20 16:27:56 +01:00
|
|
|
#include <linux/bitmap.h>
|
2017-11-28 21:16:24 +01:00
|
|
|
#include <linux/bug.h>
|
2011-11-17 03:29:17 +01:00
|
|
|
#include <linux/export.h>
|
2005-04-17 00:20:36 +02:00
|
|
|
#include <linux/idr.h>
|
2016-12-20 16:27:56 +01:00
|
|
|
#include <linux/slab.h>
|
2011-08-04 01:21:06 +02:00
|
|
|
#include <linux/spinlock.h>
|
2018-06-08 02:10:45 +02:00
|
|
|
#include <linux/xarray.h>
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2017-11-28 16:14:27 +01:00
|
|
|
/**
|
|
|
|
* idr_alloc_u32() - Allocate an ID.
|
|
|
|
* @idr: IDR handle.
|
|
|
|
* @ptr: Pointer to be associated with the new ID.
|
|
|
|
* @nextid: Pointer to an ID.
|
|
|
|
* @max: The maximum ID to allocate (inclusive).
|
|
|
|
* @gfp: Memory allocation flags.
|
|
|
|
*
|
|
|
|
* Allocates an unused ID in the range specified by @nextid and @max.
|
|
|
|
* Note that @max is inclusive whereas the @end parameter to idr_alloc()
|
2017-11-28 21:16:24 +01:00
|
|
|
* is exclusive. The new ID is assigned to @nextid before the pointer
|
|
|
|
* is inserted into the IDR, so if @nextid points into the object pointed
|
|
|
|
* to by @ptr, a concurrent lookup will not find an uninitialised ID.
|
2017-11-28 16:14:27 +01:00
|
|
|
*
|
|
|
|
* The caller should provide their own locking to ensure that two
|
|
|
|
* concurrent modifications to the IDR are not possible. Read-only
|
|
|
|
* accesses to the IDR may be done under the RCU read lock or may
|
|
|
|
* exclude simultaneous writers.
|
|
|
|
*
|
|
|
|
* Return: 0 if an ID was allocated, -ENOMEM if memory allocation failed,
|
|
|
|
* or -ENOSPC if no free IDs could be found. If an error occurred,
|
|
|
|
* @nextid is unchanged.
|
|
|
|
*/
|
|
|
|
int idr_alloc_u32(struct idr *idr, void *ptr, u32 *nextid,
|
|
|
|
unsigned long max, gfp_t gfp)
|
idr: implement idr_preload[_end]() and idr_alloc()
The current idr interface is very cumbersome.
* For all allocations, two function calls - idr_pre_get() and
idr_get_new*() - should be made.
* idr_pre_get() doesn't guarantee that the following idr_get_new*()
will not fail from memory shortage. If idr_get_new*() returns
-EAGAIN, the caller is expected to retry pre_get and allocation.
* idr_get_new*() can't enforce upper limit. Upper limit can only be
enforced by allocating and then freeing if above limit.
* idr_layer buffer is unnecessarily per-idr. Each idr ends up keeping
around MAX_IDR_FREE idr_layers. The memory consumed per idr is
under two pages but it makes it difficult to make idr_layer larger.
This patch implements the following new set of allocation functions.
* idr_preload[_end]() - Similar to radix preload but doesn't fail.
The first idr_alloc() inside preload section can be treated as if it
were called with @gfp_mask used for idr_preload().
* idr_alloc() - Allocate an ID w/ lower and upper limits. Takes
@gfp_flags and can be used w/o preloading. When used inside
preloaded section, the allocation mask of preloading can be assumed.
If idr_alloc() can be called from a context which allows sufficiently
relaxed @gfp_mask, it can be used by itself. If, for example,
idr_alloc() is called inside spinlock protected region, preloading can
be used like the following.
idr_preload(GFP_KERNEL);
spin_lock(lock);
id = idr_alloc(idr, ptr, start, end, GFP_NOWAIT);
spin_unlock(lock);
idr_preload_end();
if (id < 0)
error;
which is much simpler and less error-prone than idr_pre_get and
idr_get_new*() loop.
The new interface uses per-pcu idr_layer buffer and thus the number of
idr's in the system doesn't affect the amount of memory used for
preloading.
idr_layer_alloc() is introduced to handle idr_layer allocations for
both old and new ID allocation paths. This is a bit hairy now but the
new interface is expected to replace the old and the internal
implementation eventually will become simpler.
Signed-off-by: Tejun Heo <tj@kernel.org>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 02:03:55 +01:00
|
|
|
{
|
2016-12-20 16:27:56 +01:00
|
|
|
struct radix_tree_iter iter;
|
idr: Add new APIs to support unsigned long
The following new APIs are added:
int idr_alloc_ext(struct idr *idr, void *ptr, unsigned long *index,
unsigned long start, unsigned long end, gfp_t gfp);
void *idr_remove_ext(struct idr *idr, unsigned long id);
void *idr_find_ext(const struct idr *idr, unsigned long id);
void *idr_replace_ext(struct idr *idr, void *ptr, unsigned long id);
void *idr_get_next_ext(struct idr *idr, unsigned long *nextid);
Signed-off-by: Chris Mi <chrism@mellanox.com>
Signed-off-by: Jiri Pirko <jiri@mellanox.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-08-30 08:31:57 +02:00
|
|
|
void __rcu **slot;
|
2018-02-26 20:39:30 +01:00
|
|
|
unsigned int base = idr->idr_base;
|
|
|
|
unsigned int id = *nextid;
|
idr: implement idr_preload[_end]() and idr_alloc()
The current idr interface is very cumbersome.
* For all allocations, two function calls - idr_pre_get() and
idr_get_new*() - should be made.
* idr_pre_get() doesn't guarantee that the following idr_get_new*()
will not fail from memory shortage. If idr_get_new*() returns
-EAGAIN, the caller is expected to retry pre_get and allocation.
* idr_get_new*() can't enforce upper limit. Upper limit can only be
enforced by allocating and then freeing if above limit.
* idr_layer buffer is unnecessarily per-idr. Each idr ends up keeping
around MAX_IDR_FREE idr_layers. The memory consumed per idr is
under two pages but it makes it difficult to make idr_layer larger.
This patch implements the following new set of allocation functions.
* idr_preload[_end]() - Similar to radix preload but doesn't fail.
The first idr_alloc() inside preload section can be treated as if it
were called with @gfp_mask used for idr_preload().
* idr_alloc() - Allocate an ID w/ lower and upper limits. Takes
@gfp_flags and can be used w/o preloading. When used inside
preloaded section, the allocation mask of preloading can be assumed.
If idr_alloc() can be called from a context which allows sufficiently
relaxed @gfp_mask, it can be used by itself. If, for example,
idr_alloc() is called inside spinlock protected region, preloading can
be used like the following.
idr_preload(GFP_KERNEL);
spin_lock(lock);
id = idr_alloc(idr, ptr, start, end, GFP_NOWAIT);
spin_unlock(lock);
idr_preload_end();
if (id < 0)
error;
which is much simpler and less error-prone than idr_pre_get and
idr_get_new*() loop.
The new interface uses per-pcu idr_layer buffer and thus the number of
idr's in the system doesn't affect the amount of memory used for
preloading.
idr_layer_alloc() is introduced to handle idr_layer allocations for
both old and new ID allocation paths. This is a bit hairy now but the
new interface is expected to replace the old and the internal
implementation eventually will become simpler.
Signed-off-by: Tejun Heo <tj@kernel.org>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 02:03:55 +01:00
|
|
|
|
2017-11-07 22:30:10 +01:00
|
|
|
if (WARN_ON_ONCE(!(idr->idr_rt.xa_flags & ROOT_IS_IDR)))
|
|
|
|
idr->idr_rt.xa_flags |= IDR_RT_MARKER;
|
idr: implement idr_preload[_end]() and idr_alloc()
The current idr interface is very cumbersome.
* For all allocations, two function calls - idr_pre_get() and
idr_get_new*() - should be made.
* idr_pre_get() doesn't guarantee that the following idr_get_new*()
will not fail from memory shortage. If idr_get_new*() returns
-EAGAIN, the caller is expected to retry pre_get and allocation.
* idr_get_new*() can't enforce upper limit. Upper limit can only be
enforced by allocating and then freeing if above limit.
* idr_layer buffer is unnecessarily per-idr. Each idr ends up keeping
around MAX_IDR_FREE idr_layers. The memory consumed per idr is
under two pages but it makes it difficult to make idr_layer larger.
This patch implements the following new set of allocation functions.
* idr_preload[_end]() - Similar to radix preload but doesn't fail.
The first idr_alloc() inside preload section can be treated as if it
were called with @gfp_mask used for idr_preload().
* idr_alloc() - Allocate an ID w/ lower and upper limits. Takes
@gfp_flags and can be used w/o preloading. When used inside
preloaded section, the allocation mask of preloading can be assumed.
If idr_alloc() can be called from a context which allows sufficiently
relaxed @gfp_mask, it can be used by itself. If, for example,
idr_alloc() is called inside spinlock protected region, preloading can
be used like the following.
idr_preload(GFP_KERNEL);
spin_lock(lock);
id = idr_alloc(idr, ptr, start, end, GFP_NOWAIT);
spin_unlock(lock);
idr_preload_end();
if (id < 0)
error;
which is much simpler and less error-prone than idr_pre_get and
idr_get_new*() loop.
The new interface uses per-pcu idr_layer buffer and thus the number of
idr's in the system doesn't affect the amount of memory used for
preloading.
idr_layer_alloc() is introduced to handle idr_layer allocations for
both old and new ID allocation paths. This is a bit hairy now but the
new interface is expected to replace the old and the internal
implementation eventually will become simpler.
Signed-off-by: Tejun Heo <tj@kernel.org>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 02:03:55 +01:00
|
|
|
|
2017-11-30 19:45:11 +01:00
|
|
|
id = (id < base) ? 0 : id - base;
|
|
|
|
radix_tree_iter_init(&iter, id);
|
|
|
|
slot = idr_get_free(&idr->idr_rt, &iter, gfp, max - base);
|
2016-12-20 16:27:56 +01:00
|
|
|
if (IS_ERR(slot))
|
|
|
|
return PTR_ERR(slot);
|
idr: implement idr_preload[_end]() and idr_alloc()
The current idr interface is very cumbersome.
* For all allocations, two function calls - idr_pre_get() and
idr_get_new*() - should be made.
* idr_pre_get() doesn't guarantee that the following idr_get_new*()
will not fail from memory shortage. If idr_get_new*() returns
-EAGAIN, the caller is expected to retry pre_get and allocation.
* idr_get_new*() can't enforce upper limit. Upper limit can only be
enforced by allocating and then freeing if above limit.
* idr_layer buffer is unnecessarily per-idr. Each idr ends up keeping
around MAX_IDR_FREE idr_layers. The memory consumed per idr is
under two pages but it makes it difficult to make idr_layer larger.
This patch implements the following new set of allocation functions.
* idr_preload[_end]() - Similar to radix preload but doesn't fail.
The first idr_alloc() inside preload section can be treated as if it
were called with @gfp_mask used for idr_preload().
* idr_alloc() - Allocate an ID w/ lower and upper limits. Takes
@gfp_flags and can be used w/o preloading. When used inside
preloaded section, the allocation mask of preloading can be assumed.
If idr_alloc() can be called from a context which allows sufficiently
relaxed @gfp_mask, it can be used by itself. If, for example,
idr_alloc() is called inside spinlock protected region, preloading can
be used like the following.
idr_preload(GFP_KERNEL);
spin_lock(lock);
id = idr_alloc(idr, ptr, start, end, GFP_NOWAIT);
spin_unlock(lock);
idr_preload_end();
if (id < 0)
error;
which is much simpler and less error-prone than idr_pre_get and
idr_get_new*() loop.
The new interface uses per-pcu idr_layer buffer and thus the number of
idr's in the system doesn't affect the amount of memory used for
preloading.
idr_layer_alloc() is introduced to handle idr_layer allocations for
both old and new ID allocation paths. This is a bit hairy now but the
new interface is expected to replace the old and the internal
implementation eventually will become simpler.
Signed-off-by: Tejun Heo <tj@kernel.org>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 02:03:55 +01:00
|
|
|
|
2017-11-30 19:45:11 +01:00
|
|
|
*nextid = iter.index + base;
|
2017-11-28 21:16:24 +01:00
|
|
|
/* there is a memory barrier inside radix_tree_iter_replace() */
|
2016-12-20 16:27:56 +01:00
|
|
|
radix_tree_iter_replace(&idr->idr_rt, &iter, slot, ptr);
|
|
|
|
radix_tree_iter_tag_clear(&idr->idr_rt, &iter, IDR_FREE);
|
idr: Add new APIs to support unsigned long
The following new APIs are added:
int idr_alloc_ext(struct idr *idr, void *ptr, unsigned long *index,
unsigned long start, unsigned long end, gfp_t gfp);
void *idr_remove_ext(struct idr *idr, unsigned long id);
void *idr_find_ext(const struct idr *idr, unsigned long id);
void *idr_replace_ext(struct idr *idr, void *ptr, unsigned long id);
void *idr_get_next_ext(struct idr *idr, unsigned long *nextid);
Signed-off-by: Chris Mi <chrism@mellanox.com>
Signed-off-by: Jiri Pirko <jiri@mellanox.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-08-30 08:31:57 +02:00
|
|
|
|
|
|
|
return 0;
|
idr: implement idr_preload[_end]() and idr_alloc()
The current idr interface is very cumbersome.
* For all allocations, two function calls - idr_pre_get() and
idr_get_new*() - should be made.
* idr_pre_get() doesn't guarantee that the following idr_get_new*()
will not fail from memory shortage. If idr_get_new*() returns
-EAGAIN, the caller is expected to retry pre_get and allocation.
* idr_get_new*() can't enforce upper limit. Upper limit can only be
enforced by allocating and then freeing if above limit.
* idr_layer buffer is unnecessarily per-idr. Each idr ends up keeping
around MAX_IDR_FREE idr_layers. The memory consumed per idr is
under two pages but it makes it difficult to make idr_layer larger.
This patch implements the following new set of allocation functions.
* idr_preload[_end]() - Similar to radix preload but doesn't fail.
The first idr_alloc() inside preload section can be treated as if it
were called with @gfp_mask used for idr_preload().
* idr_alloc() - Allocate an ID w/ lower and upper limits. Takes
@gfp_flags and can be used w/o preloading. When used inside
preloaded section, the allocation mask of preloading can be assumed.
If idr_alloc() can be called from a context which allows sufficiently
relaxed @gfp_mask, it can be used by itself. If, for example,
idr_alloc() is called inside spinlock protected region, preloading can
be used like the following.
idr_preload(GFP_KERNEL);
spin_lock(lock);
id = idr_alloc(idr, ptr, start, end, GFP_NOWAIT);
spin_unlock(lock);
idr_preload_end();
if (id < 0)
error;
which is much simpler and less error-prone than idr_pre_get and
idr_get_new*() loop.
The new interface uses per-pcu idr_layer buffer and thus the number of
idr's in the system doesn't affect the amount of memory used for
preloading.
idr_layer_alloc() is introduced to handle idr_layer allocations for
both old and new ID allocation paths. This is a bit hairy now but the
new interface is expected to replace the old and the internal
implementation eventually will become simpler.
Signed-off-by: Tejun Heo <tj@kernel.org>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 02:03:55 +01:00
|
|
|
}
|
2017-11-28 21:16:24 +01:00
|
|
|
EXPORT_SYMBOL_GPL(idr_alloc_u32);
|
idr: implement idr_preload[_end]() and idr_alloc()
The current idr interface is very cumbersome.
* For all allocations, two function calls - idr_pre_get() and
idr_get_new*() - should be made.
* idr_pre_get() doesn't guarantee that the following idr_get_new*()
will not fail from memory shortage. If idr_get_new*() returns
-EAGAIN, the caller is expected to retry pre_get and allocation.
* idr_get_new*() can't enforce upper limit. Upper limit can only be
enforced by allocating and then freeing if above limit.
* idr_layer buffer is unnecessarily per-idr. Each idr ends up keeping
around MAX_IDR_FREE idr_layers. The memory consumed per idr is
under two pages but it makes it difficult to make idr_layer larger.
This patch implements the following new set of allocation functions.
* idr_preload[_end]() - Similar to radix preload but doesn't fail.
The first idr_alloc() inside preload section can be treated as if it
were called with @gfp_mask used for idr_preload().
* idr_alloc() - Allocate an ID w/ lower and upper limits. Takes
@gfp_flags and can be used w/o preloading. When used inside
preloaded section, the allocation mask of preloading can be assumed.
If idr_alloc() can be called from a context which allows sufficiently
relaxed @gfp_mask, it can be used by itself. If, for example,
idr_alloc() is called inside spinlock protected region, preloading can
be used like the following.
idr_preload(GFP_KERNEL);
spin_lock(lock);
id = idr_alloc(idr, ptr, start, end, GFP_NOWAIT);
spin_unlock(lock);
idr_preload_end();
if (id < 0)
error;
which is much simpler and less error-prone than idr_pre_get and
idr_get_new*() loop.
The new interface uses per-pcu idr_layer buffer and thus the number of
idr's in the system doesn't affect the amount of memory used for
preloading.
idr_layer_alloc() is introduced to handle idr_layer allocations for
both old and new ID allocation paths. This is a bit hairy now but the
new interface is expected to replace the old and the internal
implementation eventually will become simpler.
Signed-off-by: Tejun Heo <tj@kernel.org>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 02:03:55 +01:00
|
|
|
|
2013-04-30 01:21:16 +02:00
|
|
|
/**
|
2017-11-28 21:16:24 +01:00
|
|
|
* idr_alloc() - Allocate an ID.
|
|
|
|
* @idr: IDR handle.
|
|
|
|
* @ptr: Pointer to be associated with the new ID.
|
|
|
|
* @start: The minimum ID (inclusive).
|
|
|
|
* @end: The maximum ID (exclusive).
|
|
|
|
* @gfp: Memory allocation flags.
|
2013-02-28 02:03:35 +01:00
|
|
|
*
|
2017-11-28 21:16:24 +01:00
|
|
|
* Allocates an unused ID in the range specified by @start and @end. If
|
|
|
|
* @end is <= 0, it is treated as one larger than %INT_MAX. This allows
|
|
|
|
* callers to use @start + N as @end as long as N is within integer range.
|
|
|
|
*
|
|
|
|
* The caller should provide their own locking to ensure that two
|
|
|
|
* concurrent modifications to the IDR are not possible. Read-only
|
|
|
|
* accesses to the IDR may be done under the RCU read lock or may
|
|
|
|
* exclude simultaneous writers.
|
|
|
|
*
|
|
|
|
* Return: The newly allocated ID, -ENOMEM if memory allocation failed,
|
|
|
|
* or -ENOSPC if no free IDs could be found.
|
2005-10-23 21:57:18 +02:00
|
|
|
*/
|
2017-11-28 21:16:24 +01:00
|
|
|
int idr_alloc(struct idr *idr, void *ptr, int start, int end, gfp_t gfp)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2017-11-28 21:16:24 +01:00
|
|
|
u32 id = start;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (WARN_ON_ONCE(start < 0))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
ret = idr_alloc_u32(idr, ptr, &id, end > 0 ? end - 1 : INT_MAX, gfp);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2017-11-28 21:16:24 +01:00
|
|
|
return id;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(idr_alloc);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* idr_alloc_cyclic() - Allocate an ID cyclically.
|
|
|
|
* @idr: IDR handle.
|
|
|
|
* @ptr: Pointer to be associated with the new ID.
|
|
|
|
* @start: The minimum ID (inclusive).
|
|
|
|
* @end: The maximum ID (exclusive).
|
|
|
|
* @gfp: Memory allocation flags.
|
|
|
|
*
|
|
|
|
* Allocates an unused ID in the range specified by @nextid and @end. If
|
|
|
|
* @end is <= 0, it is treated as one larger than %INT_MAX. This allows
|
|
|
|
* callers to use @start + N as @end as long as N is within integer range.
|
|
|
|
* The search for an unused ID will start at the last ID allocated and will
|
|
|
|
* wrap around to @start if no free IDs are found before reaching @end.
|
|
|
|
*
|
|
|
|
* The caller should provide their own locking to ensure that two
|
|
|
|
* concurrent modifications to the IDR are not possible. Read-only
|
|
|
|
* accesses to the IDR may be done under the RCU read lock or may
|
|
|
|
* exclude simultaneous writers.
|
|
|
|
*
|
|
|
|
* Return: The newly allocated ID, -ENOMEM if memory allocation failed,
|
|
|
|
* or -ENOSPC if no free IDs could be found.
|
|
|
|
*/
|
|
|
|
int idr_alloc_cyclic(struct idr *idr, void *ptr, int start, int end, gfp_t gfp)
|
|
|
|
{
|
|
|
|
u32 id = idr->idr_next;
|
|
|
|
int err, max = end > 0 ? end - 1 : INT_MAX;
|
idr: remove MAX_IDR_MASK and move left MAX_IDR_* into idr.c
MAX_IDR_MASK is another weirdness in the idr interface. As idr covers
whole positive integer range, it's defined as 0x7fffffff or INT_MAX.
Its usage in idr_find(), idr_replace() and idr_remove() is bizarre.
They basically mask off the sign bit and operate on the rest, so if
the caller, by accident, passes in a negative number, the sign bit
will be masked off and the remaining part will be used as if that was
the input, which is worse than crashing.
The constant is visible in idr.h and there are several users in the
kernel.
* drivers/i2c/i2c-core.c:i2c_add_numbered_adapter()
Basically used to test if adap->nr is a negative number which isn't
-1 and returns -EINVAL if so. idr_alloc() already has negative
@start checking (w/ WARN_ON_ONCE), so this can go away.
* drivers/infiniband/core/cm.c:cm_alloc_id()
drivers/infiniband/hw/mlx4/cm.c:id_map_alloc()
Used to wrap cyclic @start. Can be replaced with max(next, 0).
Note that this type of cyclic allocation using idr is buggy. These
are prone to spurious -ENOSPC failure after the first wraparound.
* fs/super.c:get_anon_bdev()
The ID allocated from ida is masked off before being tested whether
it's inside valid range. ida allocated ID can never be a negative
number and the masking is unnecessary.
Update idr_*() functions to fail with -EINVAL when negative @id is
specified and update other MAX_IDR_MASK users as described above.
This leaves MAX_IDR_MASK without any user, remove it and relocate
other MAX_IDR_* constants to lib/idr.c.
Signed-off-by: Tejun Heo <tj@kernel.org>
Cc: Jean Delvare <khali@linux-fr.org>
Cc: Roland Dreier <roland@kernel.org>
Cc: Sean Hefty <sean.hefty@intel.com>
Cc: Hal Rosenstock <hal.rosenstock@gmail.com>
Cc: "Marciniszyn, Mike" <mike.marciniszyn@intel.com>
Cc: Jack Morgenstein <jackm@dev.mellanox.co.il>
Cc: Or Gerlitz <ogerlitz@mellanox.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Acked-by: Wolfram Sang <wolfram@the-dreams.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 02:05:04 +01:00
|
|
|
|
2017-11-28 21:16:24 +01:00
|
|
|
if ((int)id < start)
|
|
|
|
id = start;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2017-11-28 21:16:24 +01:00
|
|
|
err = idr_alloc_u32(idr, ptr, &id, max, gfp);
|
|
|
|
if ((err == -ENOSPC) && (id > start)) {
|
|
|
|
id = start;
|
|
|
|
err = idr_alloc_u32(idr, ptr, &id, max, gfp);
|
|
|
|
}
|
|
|
|
if (err)
|
|
|
|
return err;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2017-11-28 21:16:24 +01:00
|
|
|
idr->idr_next = id + 1;
|
2016-12-20 16:27:56 +01:00
|
|
|
return id;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
2016-12-20 16:27:56 +01:00
|
|
|
EXPORT_SYMBOL(idr_alloc_cyclic);
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2017-11-30 19:45:11 +01:00
|
|
|
/**
|
|
|
|
* idr_remove() - Remove an ID from the IDR.
|
|
|
|
* @idr: IDR handle.
|
|
|
|
* @id: Pointer ID.
|
|
|
|
*
|
|
|
|
* Removes this ID from the IDR. If the ID was not previously in the IDR,
|
|
|
|
* this function returns %NULL.
|
|
|
|
*
|
|
|
|
* Since this function modifies the IDR, the caller should provide their
|
|
|
|
* own locking to ensure that concurrent modification of the same IDR is
|
|
|
|
* not possible.
|
|
|
|
*
|
|
|
|
* Return: The pointer formerly associated with this ID.
|
|
|
|
*/
|
|
|
|
void *idr_remove(struct idr *idr, unsigned long id)
|
|
|
|
{
|
|
|
|
return radix_tree_delete_item(&idr->idr_rt, id - idr->idr_base, NULL);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(idr_remove);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* idr_find() - Return pointer for given ID.
|
|
|
|
* @idr: IDR handle.
|
|
|
|
* @id: Pointer ID.
|
|
|
|
*
|
|
|
|
* Looks up the pointer associated with this ID. A %NULL pointer may
|
|
|
|
* indicate that @id is not allocated or that the %NULL pointer was
|
|
|
|
* associated with this ID.
|
|
|
|
*
|
|
|
|
* This function can be called under rcu_read_lock(), given that the leaf
|
|
|
|
* pointers lifetimes are correctly managed.
|
|
|
|
*
|
|
|
|
* Return: The pointer associated with this ID.
|
|
|
|
*/
|
|
|
|
void *idr_find(const struct idr *idr, unsigned long id)
|
|
|
|
{
|
|
|
|
return radix_tree_lookup(&idr->idr_rt, id - idr->idr_base);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(idr_find);
|
|
|
|
|
2007-07-16 08:37:24 +02:00
|
|
|
/**
|
2017-11-28 21:39:51 +01:00
|
|
|
* idr_for_each() - Iterate through all stored pointers.
|
|
|
|
* @idr: IDR handle.
|
|
|
|
* @fn: Function to be called for each pointer.
|
|
|
|
* @data: Data passed to callback function.
|
2007-07-16 08:37:24 +02:00
|
|
|
*
|
2016-12-20 16:27:56 +01:00
|
|
|
* The callback function will be called for each entry in @idr, passing
|
2017-11-28 21:39:51 +01:00
|
|
|
* the ID, the entry and @data.
|
2007-07-16 08:37:24 +02:00
|
|
|
*
|
2016-12-20 16:27:56 +01:00
|
|
|
* If @fn returns anything other than %0, the iteration stops and that
|
|
|
|
* value is returned from this function.
|
2007-07-16 08:37:24 +02:00
|
|
|
*
|
2016-12-20 16:27:56 +01:00
|
|
|
* idr_for_each() can be called concurrently with idr_alloc() and
|
|
|
|
* idr_remove() if protected by RCU. Newly added entries may not be
|
|
|
|
* seen and deleted entries may be seen, but adding and removing entries
|
|
|
|
* will not cause other entries to be skipped, nor spurious ones to be seen.
|
2007-07-16 08:37:24 +02:00
|
|
|
*/
|
2016-12-20 16:27:56 +01:00
|
|
|
int idr_for_each(const struct idr *idr,
|
|
|
|
int (*fn)(int id, void *p, void *data), void *data)
|
2007-07-16 08:37:24 +02:00
|
|
|
{
|
2016-12-20 16:27:56 +01:00
|
|
|
struct radix_tree_iter iter;
|
2017-02-13 22:03:55 +01:00
|
|
|
void __rcu **slot;
|
2017-11-30 19:45:11 +01:00
|
|
|
int base = idr->idr_base;
|
2007-07-16 08:37:24 +02:00
|
|
|
|
2016-12-20 16:27:56 +01:00
|
|
|
radix_tree_for_each_slot(slot, &idr->idr_rt, &iter, 0) {
|
2017-11-28 21:50:12 +01:00
|
|
|
int ret;
|
2018-02-26 20:39:30 +01:00
|
|
|
unsigned long id = iter.index + base;
|
2017-11-28 21:50:12 +01:00
|
|
|
|
2018-02-26 20:39:30 +01:00
|
|
|
if (WARN_ON_ONCE(id > INT_MAX))
|
2017-11-28 21:50:12 +01:00
|
|
|
break;
|
2018-02-26 20:39:30 +01:00
|
|
|
ret = fn(id, rcu_dereference_raw(*slot), data);
|
2016-12-20 16:27:56 +01:00
|
|
|
if (ret)
|
|
|
|
return ret;
|
2007-07-16 08:37:24 +02:00
|
|
|
}
|
|
|
|
|
2016-12-20 16:27:56 +01:00
|
|
|
return 0;
|
2007-07-16 08:37:24 +02:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(idr_for_each);
|
|
|
|
|
cgroup: CSS ID support
Patch for Per-CSS(Cgroup Subsys State) ID and private hierarchy code.
This patch attaches unique ID to each css and provides following.
- css_lookup(subsys, id)
returns pointer to struct cgroup_subysys_state of id.
- css_get_next(subsys, id, rootid, depth, foundid)
returns the next css under "root" by scanning
When cgroup_subsys->use_id is set, an id for css is maintained.
The cgroup framework only parepares
- css_id of root css for subsys
- id is automatically attached at creation of css.
- id is *not* freed automatically. Because the cgroup framework
don't know lifetime of cgroup_subsys_state.
free_css_id() function is provided. This must be called by subsys.
There are several reasons to develop this.
- Saving space .... For example, memcg's swap_cgroup is array of
pointers to cgroup. But it is not necessary to be very fast.
By replacing pointers(8bytes per ent) to ID (2byes per ent), we can
reduce much amount of memory usage.
- Scanning without lock.
CSS_ID provides "scan id under this ROOT" function. By this, scanning
css under root can be written without locks.
ex)
do {
rcu_read_lock();
next = cgroup_get_next(subsys, id, root, &found);
/* check sanity of next here */
css_tryget();
rcu_read_unlock();
id = found + 1
} while(...)
Characteristics:
- Each css has unique ID under subsys.
- Lifetime of ID is controlled by subsys.
- css ID contains "ID" and "Depth in hierarchy" and stack of hierarchy
- Allowed ID is 1-65535, ID 0 is UNUSED ID.
Design Choices:
- scan-by-ID v.s. scan-by-tree-walk.
As /proc's pid scan does, scan-by-ID is robust when scanning is done
by following kind of routine.
scan -> rest a while(release a lock) -> conitunue from interrupted
memcg's hierarchical reclaim does this.
- When subsys->use_id is set, # of css in the system is limited to
65535.
[bharata@linux.vnet.ibm.com: remove rcu_read_lock() from css_get_next()]
Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Acked-by: Paul Menage <menage@google.com>
Cc: Li Zefan <lizf@cn.fujitsu.com>
Cc: Balbir Singh <balbir@in.ibm.com>
Cc: Daisuke Nishimura <nishimura@mxp.nes.nec.co.jp>
Signed-off-by: Bharata B Rao <bharata@linux.vnet.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-03 01:57:25 +02:00
|
|
|
/**
|
2019-11-02 02:36:39 +01:00
|
|
|
* idr_get_next_ul() - Find next populated entry.
|
2017-11-28 21:39:51 +01:00
|
|
|
* @idr: IDR handle.
|
|
|
|
* @nextid: Pointer to an ID.
|
2016-12-20 16:27:56 +01:00
|
|
|
*
|
|
|
|
* Returns the next populated entry in the tree with an ID greater than
|
|
|
|
* or equal to the value pointed to by @nextid. On exit, @nextid is updated
|
|
|
|
* to the ID of the found value. To use in a loop, the value pointed to by
|
|
|
|
* nextid must be incremented by the user.
|
cgroup: CSS ID support
Patch for Per-CSS(Cgroup Subsys State) ID and private hierarchy code.
This patch attaches unique ID to each css and provides following.
- css_lookup(subsys, id)
returns pointer to struct cgroup_subysys_state of id.
- css_get_next(subsys, id, rootid, depth, foundid)
returns the next css under "root" by scanning
When cgroup_subsys->use_id is set, an id for css is maintained.
The cgroup framework only parepares
- css_id of root css for subsys
- id is automatically attached at creation of css.
- id is *not* freed automatically. Because the cgroup framework
don't know lifetime of cgroup_subsys_state.
free_css_id() function is provided. This must be called by subsys.
There are several reasons to develop this.
- Saving space .... For example, memcg's swap_cgroup is array of
pointers to cgroup. But it is not necessary to be very fast.
By replacing pointers(8bytes per ent) to ID (2byes per ent), we can
reduce much amount of memory usage.
- Scanning without lock.
CSS_ID provides "scan id under this ROOT" function. By this, scanning
css under root can be written without locks.
ex)
do {
rcu_read_lock();
next = cgroup_get_next(subsys, id, root, &found);
/* check sanity of next here */
css_tryget();
rcu_read_unlock();
id = found + 1
} while(...)
Characteristics:
- Each css has unique ID under subsys.
- Lifetime of ID is controlled by subsys.
- css ID contains "ID" and "Depth in hierarchy" and stack of hierarchy
- Allowed ID is 1-65535, ID 0 is UNUSED ID.
Design Choices:
- scan-by-ID v.s. scan-by-tree-walk.
As /proc's pid scan does, scan-by-ID is robust when scanning is done
by following kind of routine.
scan -> rest a while(release a lock) -> conitunue from interrupted
memcg's hierarchical reclaim does this.
- When subsys->use_id is set, # of css in the system is limited to
65535.
[bharata@linux.vnet.ibm.com: remove rcu_read_lock() from css_get_next()]
Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Acked-by: Paul Menage <menage@google.com>
Cc: Li Zefan <lizf@cn.fujitsu.com>
Cc: Balbir Singh <balbir@in.ibm.com>
Cc: Daisuke Nishimura <nishimura@mxp.nes.nec.co.jp>
Signed-off-by: Bharata B Rao <bharata@linux.vnet.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-03 01:57:25 +02:00
|
|
|
*/
|
2019-11-02 02:36:39 +01:00
|
|
|
void *idr_get_next_ul(struct idr *idr, unsigned long *nextid)
|
cgroup: CSS ID support
Patch for Per-CSS(Cgroup Subsys State) ID and private hierarchy code.
This patch attaches unique ID to each css and provides following.
- css_lookup(subsys, id)
returns pointer to struct cgroup_subysys_state of id.
- css_get_next(subsys, id, rootid, depth, foundid)
returns the next css under "root" by scanning
When cgroup_subsys->use_id is set, an id for css is maintained.
The cgroup framework only parepares
- css_id of root css for subsys
- id is automatically attached at creation of css.
- id is *not* freed automatically. Because the cgroup framework
don't know lifetime of cgroup_subsys_state.
free_css_id() function is provided. This must be called by subsys.
There are several reasons to develop this.
- Saving space .... For example, memcg's swap_cgroup is array of
pointers to cgroup. But it is not necessary to be very fast.
By replacing pointers(8bytes per ent) to ID (2byes per ent), we can
reduce much amount of memory usage.
- Scanning without lock.
CSS_ID provides "scan id under this ROOT" function. By this, scanning
css under root can be written without locks.
ex)
do {
rcu_read_lock();
next = cgroup_get_next(subsys, id, root, &found);
/* check sanity of next here */
css_tryget();
rcu_read_unlock();
id = found + 1
} while(...)
Characteristics:
- Each css has unique ID under subsys.
- Lifetime of ID is controlled by subsys.
- css ID contains "ID" and "Depth in hierarchy" and stack of hierarchy
- Allowed ID is 1-65535, ID 0 is UNUSED ID.
Design Choices:
- scan-by-ID v.s. scan-by-tree-walk.
As /proc's pid scan does, scan-by-ID is robust when scanning is done
by following kind of routine.
scan -> rest a while(release a lock) -> conitunue from interrupted
memcg's hierarchical reclaim does this.
- When subsys->use_id is set, # of css in the system is limited to
65535.
[bharata@linux.vnet.ibm.com: remove rcu_read_lock() from css_get_next()]
Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Acked-by: Paul Menage <menage@google.com>
Cc: Li Zefan <lizf@cn.fujitsu.com>
Cc: Balbir Singh <balbir@in.ibm.com>
Cc: Daisuke Nishimura <nishimura@mxp.nes.nec.co.jp>
Signed-off-by: Bharata B Rao <bharata@linux.vnet.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-03 01:57:25 +02:00
|
|
|
{
|
2016-12-20 16:27:56 +01:00
|
|
|
struct radix_tree_iter iter;
|
2017-02-13 22:03:55 +01:00
|
|
|
void __rcu **slot;
|
2019-05-14 22:05:45 +02:00
|
|
|
void *entry = NULL;
|
2018-02-26 20:39:30 +01:00
|
|
|
unsigned long base = idr->idr_base;
|
|
|
|
unsigned long id = *nextid;
|
cgroup: CSS ID support
Patch for Per-CSS(Cgroup Subsys State) ID and private hierarchy code.
This patch attaches unique ID to each css and provides following.
- css_lookup(subsys, id)
returns pointer to struct cgroup_subysys_state of id.
- css_get_next(subsys, id, rootid, depth, foundid)
returns the next css under "root" by scanning
When cgroup_subsys->use_id is set, an id for css is maintained.
The cgroup framework only parepares
- css_id of root css for subsys
- id is automatically attached at creation of css.
- id is *not* freed automatically. Because the cgroup framework
don't know lifetime of cgroup_subsys_state.
free_css_id() function is provided. This must be called by subsys.
There are several reasons to develop this.
- Saving space .... For example, memcg's swap_cgroup is array of
pointers to cgroup. But it is not necessary to be very fast.
By replacing pointers(8bytes per ent) to ID (2byes per ent), we can
reduce much amount of memory usage.
- Scanning without lock.
CSS_ID provides "scan id under this ROOT" function. By this, scanning
css under root can be written without locks.
ex)
do {
rcu_read_lock();
next = cgroup_get_next(subsys, id, root, &found);
/* check sanity of next here */
css_tryget();
rcu_read_unlock();
id = found + 1
} while(...)
Characteristics:
- Each css has unique ID under subsys.
- Lifetime of ID is controlled by subsys.
- css ID contains "ID" and "Depth in hierarchy" and stack of hierarchy
- Allowed ID is 1-65535, ID 0 is UNUSED ID.
Design Choices:
- scan-by-ID v.s. scan-by-tree-walk.
As /proc's pid scan does, scan-by-ID is robust when scanning is done
by following kind of routine.
scan -> rest a while(release a lock) -> conitunue from interrupted
memcg's hierarchical reclaim does this.
- When subsys->use_id is set, # of css in the system is limited to
65535.
[bharata@linux.vnet.ibm.com: remove rcu_read_lock() from css_get_next()]
Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Acked-by: Paul Menage <menage@google.com>
Cc: Li Zefan <lizf@cn.fujitsu.com>
Cc: Balbir Singh <balbir@in.ibm.com>
Cc: Daisuke Nishimura <nishimura@mxp.nes.nec.co.jp>
Signed-off-by: Bharata B Rao <bharata@linux.vnet.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-03 01:57:25 +02:00
|
|
|
|
2017-11-30 19:45:11 +01:00
|
|
|
id = (id < base) ? 0 : id - base;
|
2019-05-14 22:05:45 +02:00
|
|
|
radix_tree_for_each_slot(slot, &idr->idr_rt, &iter, id) {
|
|
|
|
entry = rcu_dereference_raw(*slot);
|
|
|
|
if (!entry)
|
|
|
|
continue;
|
|
|
|
if (!xa_is_internal(entry))
|
|
|
|
break;
|
|
|
|
if (slot != &idr->idr_rt.xa_head && !xa_is_retry(entry))
|
|
|
|
break;
|
|
|
|
slot = radix_tree_iter_retry(&iter);
|
|
|
|
}
|
2016-12-20 16:27:56 +01:00
|
|
|
if (!slot)
|
cgroup: CSS ID support
Patch for Per-CSS(Cgroup Subsys State) ID and private hierarchy code.
This patch attaches unique ID to each css and provides following.
- css_lookup(subsys, id)
returns pointer to struct cgroup_subysys_state of id.
- css_get_next(subsys, id, rootid, depth, foundid)
returns the next css under "root" by scanning
When cgroup_subsys->use_id is set, an id for css is maintained.
The cgroup framework only parepares
- css_id of root css for subsys
- id is automatically attached at creation of css.
- id is *not* freed automatically. Because the cgroup framework
don't know lifetime of cgroup_subsys_state.
free_css_id() function is provided. This must be called by subsys.
There are several reasons to develop this.
- Saving space .... For example, memcg's swap_cgroup is array of
pointers to cgroup. But it is not necessary to be very fast.
By replacing pointers(8bytes per ent) to ID (2byes per ent), we can
reduce much amount of memory usage.
- Scanning without lock.
CSS_ID provides "scan id under this ROOT" function. By this, scanning
css under root can be written without locks.
ex)
do {
rcu_read_lock();
next = cgroup_get_next(subsys, id, root, &found);
/* check sanity of next here */
css_tryget();
rcu_read_unlock();
id = found + 1
} while(...)
Characteristics:
- Each css has unique ID under subsys.
- Lifetime of ID is controlled by subsys.
- css ID contains "ID" and "Depth in hierarchy" and stack of hierarchy
- Allowed ID is 1-65535, ID 0 is UNUSED ID.
Design Choices:
- scan-by-ID v.s. scan-by-tree-walk.
As /proc's pid scan does, scan-by-ID is robust when scanning is done
by following kind of routine.
scan -> rest a while(release a lock) -> conitunue from interrupted
memcg's hierarchical reclaim does this.
- When subsys->use_id is set, # of css in the system is limited to
65535.
[bharata@linux.vnet.ibm.com: remove rcu_read_lock() from css_get_next()]
Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Acked-by: Paul Menage <menage@google.com>
Cc: Li Zefan <lizf@cn.fujitsu.com>
Cc: Balbir Singh <balbir@in.ibm.com>
Cc: Daisuke Nishimura <nishimura@mxp.nes.nec.co.jp>
Signed-off-by: Bharata B Rao <bharata@linux.vnet.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-03 01:57:25 +02:00
|
|
|
return NULL;
|
2017-11-28 21:50:12 +01:00
|
|
|
|
2019-11-02 02:36:39 +01:00
|
|
|
*nextid = iter.index + base;
|
2019-05-14 22:05:45 +02:00
|
|
|
return entry;
|
cgroup: CSS ID support
Patch for Per-CSS(Cgroup Subsys State) ID and private hierarchy code.
This patch attaches unique ID to each css and provides following.
- css_lookup(subsys, id)
returns pointer to struct cgroup_subysys_state of id.
- css_get_next(subsys, id, rootid, depth, foundid)
returns the next css under "root" by scanning
When cgroup_subsys->use_id is set, an id for css is maintained.
The cgroup framework only parepares
- css_id of root css for subsys
- id is automatically attached at creation of css.
- id is *not* freed automatically. Because the cgroup framework
don't know lifetime of cgroup_subsys_state.
free_css_id() function is provided. This must be called by subsys.
There are several reasons to develop this.
- Saving space .... For example, memcg's swap_cgroup is array of
pointers to cgroup. But it is not necessary to be very fast.
By replacing pointers(8bytes per ent) to ID (2byes per ent), we can
reduce much amount of memory usage.
- Scanning without lock.
CSS_ID provides "scan id under this ROOT" function. By this, scanning
css under root can be written without locks.
ex)
do {
rcu_read_lock();
next = cgroup_get_next(subsys, id, root, &found);
/* check sanity of next here */
css_tryget();
rcu_read_unlock();
id = found + 1
} while(...)
Characteristics:
- Each css has unique ID under subsys.
- Lifetime of ID is controlled by subsys.
- css ID contains "ID" and "Depth in hierarchy" and stack of hierarchy
- Allowed ID is 1-65535, ID 0 is UNUSED ID.
Design Choices:
- scan-by-ID v.s. scan-by-tree-walk.
As /proc's pid scan does, scan-by-ID is robust when scanning is done
by following kind of routine.
scan -> rest a while(release a lock) -> conitunue from interrupted
memcg's hierarchical reclaim does this.
- When subsys->use_id is set, # of css in the system is limited to
65535.
[bharata@linux.vnet.ibm.com: remove rcu_read_lock() from css_get_next()]
Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Acked-by: Paul Menage <menage@google.com>
Cc: Li Zefan <lizf@cn.fujitsu.com>
Cc: Balbir Singh <balbir@in.ibm.com>
Cc: Daisuke Nishimura <nishimura@mxp.nes.nec.co.jp>
Signed-off-by: Bharata B Rao <bharata@linux.vnet.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-03 01:57:25 +02:00
|
|
|
}
|
2019-11-02 02:36:39 +01:00
|
|
|
EXPORT_SYMBOL(idr_get_next_ul);
|
cgroup: CSS ID support
Patch for Per-CSS(Cgroup Subsys State) ID and private hierarchy code.
This patch attaches unique ID to each css and provides following.
- css_lookup(subsys, id)
returns pointer to struct cgroup_subysys_state of id.
- css_get_next(subsys, id, rootid, depth, foundid)
returns the next css under "root" by scanning
When cgroup_subsys->use_id is set, an id for css is maintained.
The cgroup framework only parepares
- css_id of root css for subsys
- id is automatically attached at creation of css.
- id is *not* freed automatically. Because the cgroup framework
don't know lifetime of cgroup_subsys_state.
free_css_id() function is provided. This must be called by subsys.
There are several reasons to develop this.
- Saving space .... For example, memcg's swap_cgroup is array of
pointers to cgroup. But it is not necessary to be very fast.
By replacing pointers(8bytes per ent) to ID (2byes per ent), we can
reduce much amount of memory usage.
- Scanning without lock.
CSS_ID provides "scan id under this ROOT" function. By this, scanning
css under root can be written without locks.
ex)
do {
rcu_read_lock();
next = cgroup_get_next(subsys, id, root, &found);
/* check sanity of next here */
css_tryget();
rcu_read_unlock();
id = found + 1
} while(...)
Characteristics:
- Each css has unique ID under subsys.
- Lifetime of ID is controlled by subsys.
- css ID contains "ID" and "Depth in hierarchy" and stack of hierarchy
- Allowed ID is 1-65535, ID 0 is UNUSED ID.
Design Choices:
- scan-by-ID v.s. scan-by-tree-walk.
As /proc's pid scan does, scan-by-ID is robust when scanning is done
by following kind of routine.
scan -> rest a while(release a lock) -> conitunue from interrupted
memcg's hierarchical reclaim does this.
- When subsys->use_id is set, # of css in the system is limited to
65535.
[bharata@linux.vnet.ibm.com: remove rcu_read_lock() from css_get_next()]
Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Acked-by: Paul Menage <menage@google.com>
Cc: Li Zefan <lizf@cn.fujitsu.com>
Cc: Balbir Singh <balbir@in.ibm.com>
Cc: Daisuke Nishimura <nishimura@mxp.nes.nec.co.jp>
Signed-off-by: Bharata B Rao <bharata@linux.vnet.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-03 01:57:25 +02:00
|
|
|
|
2017-11-28 21:39:51 +01:00
|
|
|
/**
|
2019-11-02 02:36:39 +01:00
|
|
|
* idr_get_next() - Find next populated entry.
|
2017-11-28 21:39:51 +01:00
|
|
|
* @idr: IDR handle.
|
|
|
|
* @nextid: Pointer to an ID.
|
|
|
|
*
|
|
|
|
* Returns the next populated entry in the tree with an ID greater than
|
|
|
|
* or equal to the value pointed to by @nextid. On exit, @nextid is updated
|
|
|
|
* to the ID of the found value. To use in a loop, the value pointed to by
|
|
|
|
* nextid must be incremented by the user.
|
|
|
|
*/
|
2019-11-02 02:36:39 +01:00
|
|
|
void *idr_get_next(struct idr *idr, int *nextid)
|
idr: Add new APIs to support unsigned long
The following new APIs are added:
int idr_alloc_ext(struct idr *idr, void *ptr, unsigned long *index,
unsigned long start, unsigned long end, gfp_t gfp);
void *idr_remove_ext(struct idr *idr, unsigned long id);
void *idr_find_ext(const struct idr *idr, unsigned long id);
void *idr_replace_ext(struct idr *idr, void *ptr, unsigned long id);
void *idr_get_next_ext(struct idr *idr, unsigned long *nextid);
Signed-off-by: Chris Mi <chrism@mellanox.com>
Signed-off-by: Jiri Pirko <jiri@mellanox.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-08-30 08:31:57 +02:00
|
|
|
{
|
2017-11-30 19:45:11 +01:00
|
|
|
unsigned long id = *nextid;
|
2019-11-02 02:36:39 +01:00
|
|
|
void *entry = idr_get_next_ul(idr, &id);
|
idr: Add new APIs to support unsigned long
The following new APIs are added:
int idr_alloc_ext(struct idr *idr, void *ptr, unsigned long *index,
unsigned long start, unsigned long end, gfp_t gfp);
void *idr_remove_ext(struct idr *idr, unsigned long id);
void *idr_find_ext(const struct idr *idr, unsigned long id);
void *idr_replace_ext(struct idr *idr, void *ptr, unsigned long id);
void *idr_get_next_ext(struct idr *idr, unsigned long *nextid);
Signed-off-by: Chris Mi <chrism@mellanox.com>
Signed-off-by: Jiri Pirko <jiri@mellanox.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-08-30 08:31:57 +02:00
|
|
|
|
2019-11-02 02:36:39 +01:00
|
|
|
if (WARN_ON_ONCE(id > INT_MAX))
|
idr: Add new APIs to support unsigned long
The following new APIs are added:
int idr_alloc_ext(struct idr *idr, void *ptr, unsigned long *index,
unsigned long start, unsigned long end, gfp_t gfp);
void *idr_remove_ext(struct idr *idr, unsigned long id);
void *idr_find_ext(const struct idr *idr, unsigned long id);
void *idr_replace_ext(struct idr *idr, void *ptr, unsigned long id);
void *idr_get_next_ext(struct idr *idr, unsigned long *nextid);
Signed-off-by: Chris Mi <chrism@mellanox.com>
Signed-off-by: Jiri Pirko <jiri@mellanox.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-08-30 08:31:57 +02:00
|
|
|
return NULL;
|
2019-11-02 02:36:39 +01:00
|
|
|
*nextid = id;
|
|
|
|
return entry;
|
idr: Add new APIs to support unsigned long
The following new APIs are added:
int idr_alloc_ext(struct idr *idr, void *ptr, unsigned long *index,
unsigned long start, unsigned long end, gfp_t gfp);
void *idr_remove_ext(struct idr *idr, unsigned long id);
void *idr_find_ext(const struct idr *idr, unsigned long id);
void *idr_replace_ext(struct idr *idr, void *ptr, unsigned long id);
void *idr_get_next_ext(struct idr *idr, unsigned long *nextid);
Signed-off-by: Chris Mi <chrism@mellanox.com>
Signed-off-by: Jiri Pirko <jiri@mellanox.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-08-30 08:31:57 +02:00
|
|
|
}
|
2019-11-02 02:36:39 +01:00
|
|
|
EXPORT_SYMBOL(idr_get_next);
|
idr: Add new APIs to support unsigned long
The following new APIs are added:
int idr_alloc_ext(struct idr *idr, void *ptr, unsigned long *index,
unsigned long start, unsigned long end, gfp_t gfp);
void *idr_remove_ext(struct idr *idr, unsigned long id);
void *idr_find_ext(const struct idr *idr, unsigned long id);
void *idr_replace_ext(struct idr *idr, void *ptr, unsigned long id);
void *idr_get_next_ext(struct idr *idr, unsigned long *nextid);
Signed-off-by: Chris Mi <chrism@mellanox.com>
Signed-off-by: Jiri Pirko <jiri@mellanox.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-08-30 08:31:57 +02:00
|
|
|
|
2006-06-26 09:27:19 +02:00
|
|
|
/**
|
2017-11-28 21:16:24 +01:00
|
|
|
* idr_replace() - replace pointer for given ID.
|
|
|
|
* @idr: IDR handle.
|
|
|
|
* @ptr: New pointer to associate with the ID.
|
|
|
|
* @id: ID to change.
|
2006-06-26 09:27:19 +02:00
|
|
|
*
|
2016-12-20 16:27:56 +01:00
|
|
|
* Replace the pointer registered with an ID and return the old value.
|
|
|
|
* This function can be called under the RCU read lock concurrently with
|
|
|
|
* idr_alloc() and idr_remove() (as long as the ID being removed is not
|
|
|
|
* the one being replaced!).
|
2006-06-26 09:27:19 +02:00
|
|
|
*
|
2017-10-04 01:16:13 +02:00
|
|
|
* Returns: the old value on success. %-ENOENT indicates that @id was not
|
2017-11-28 15:56:36 +01:00
|
|
|
* found. %-EINVAL indicates that @ptr was not valid.
|
2006-06-26 09:27:19 +02:00
|
|
|
*/
|
2017-11-28 15:56:36 +01:00
|
|
|
void *idr_replace(struct idr *idr, void *ptr, unsigned long id)
|
2006-06-26 09:27:19 +02:00
|
|
|
{
|
2016-12-20 16:27:56 +01:00
|
|
|
struct radix_tree_node *node;
|
2017-02-13 22:03:55 +01:00
|
|
|
void __rcu **slot = NULL;
|
2016-12-20 16:27:56 +01:00
|
|
|
void *entry;
|
2006-06-26 09:27:19 +02:00
|
|
|
|
2017-11-30 19:45:11 +01:00
|
|
|
id -= idr->idr_base;
|
idr: remove MAX_IDR_MASK and move left MAX_IDR_* into idr.c
MAX_IDR_MASK is another weirdness in the idr interface. As idr covers
whole positive integer range, it's defined as 0x7fffffff or INT_MAX.
Its usage in idr_find(), idr_replace() and idr_remove() is bizarre.
They basically mask off the sign bit and operate on the rest, so if
the caller, by accident, passes in a negative number, the sign bit
will be masked off and the remaining part will be used as if that was
the input, which is worse than crashing.
The constant is visible in idr.h and there are several users in the
kernel.
* drivers/i2c/i2c-core.c:i2c_add_numbered_adapter()
Basically used to test if adap->nr is a negative number which isn't
-1 and returns -EINVAL if so. idr_alloc() already has negative
@start checking (w/ WARN_ON_ONCE), so this can go away.
* drivers/infiniband/core/cm.c:cm_alloc_id()
drivers/infiniband/hw/mlx4/cm.c:id_map_alloc()
Used to wrap cyclic @start. Can be replaced with max(next, 0).
Note that this type of cyclic allocation using idr is buggy. These
are prone to spurious -ENOSPC failure after the first wraparound.
* fs/super.c:get_anon_bdev()
The ID allocated from ida is masked off before being tested whether
it's inside valid range. ida allocated ID can never be a negative
number and the masking is unnecessary.
Update idr_*() functions to fail with -EINVAL when negative @id is
specified and update other MAX_IDR_MASK users as described above.
This leaves MAX_IDR_MASK without any user, remove it and relocate
other MAX_IDR_* constants to lib/idr.c.
Signed-off-by: Tejun Heo <tj@kernel.org>
Cc: Jean Delvare <khali@linux-fr.org>
Cc: Roland Dreier <roland@kernel.org>
Cc: Sean Hefty <sean.hefty@intel.com>
Cc: Hal Rosenstock <hal.rosenstock@gmail.com>
Cc: "Marciniszyn, Mike" <mike.marciniszyn@intel.com>
Cc: Jack Morgenstein <jackm@dev.mellanox.co.il>
Cc: Or Gerlitz <ogerlitz@mellanox.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Acked-by: Wolfram Sang <wolfram@the-dreams.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 02:05:04 +01:00
|
|
|
|
2016-12-20 16:27:56 +01:00
|
|
|
entry = __radix_tree_lookup(&idr->idr_rt, id, &node, &slot);
|
|
|
|
if (!slot || radix_tree_tag_get(&idr->idr_rt, id, IDR_FREE))
|
2006-06-26 09:27:19 +02:00
|
|
|
return ERR_PTR(-ENOENT);
|
|
|
|
|
2018-04-09 22:24:45 +02:00
|
|
|
__radix_tree_replace(&idr->idr_rt, node, slot, ptr);
|
2006-06-26 09:27:19 +02:00
|
|
|
|
2016-12-20 16:27:56 +01:00
|
|
|
return entry;
|
2006-06-26 09:27:19 +02:00
|
|
|
}
|
2017-11-28 15:56:36 +01:00
|
|
|
EXPORT_SYMBOL(idr_replace);
|
2006-06-26 09:27:19 +02:00
|
|
|
|
2010-10-26 23:19:08 +02:00
|
|
|
/**
|
|
|
|
* DOC: IDA description
|
2007-06-13 20:45:13 +02:00
|
|
|
*
|
2016-12-20 16:27:56 +01:00
|
|
|
* The IDA is an ID allocator which does not provide the ability to
|
|
|
|
* associate an ID with a pointer. As such, it only needs to store one
|
|
|
|
* bit per ID, and so is more space efficient than an IDR. To use an IDA,
|
|
|
|
* define it using DEFINE_IDA() (or embed a &struct ida in a data structure,
|
|
|
|
* then initialise it using ida_init()). To allocate a new ID, call
|
2018-03-20 22:07:11 +01:00
|
|
|
* ida_alloc(), ida_alloc_min(), ida_alloc_max() or ida_alloc_range().
|
|
|
|
* To free an ID, call ida_free().
|
2016-12-20 16:27:56 +01:00
|
|
|
*
|
2018-06-19 01:02:48 +02:00
|
|
|
* ida_destroy() can be used to dispose of an IDA without needing to
|
|
|
|
* free the individual IDs in it. You can use ida_is_empty() to find
|
|
|
|
* out whether the IDA has any IDs currently allocated.
|
2016-12-20 16:27:56 +01:00
|
|
|
*
|
2018-07-04 21:42:46 +02:00
|
|
|
* The IDA handles its own locking. It is safe to call any of the IDA
|
|
|
|
* functions without synchronisation in your code.
|
|
|
|
*
|
2016-12-20 16:27:56 +01:00
|
|
|
* IDs are currently limited to the range [0-INT_MAX]. If this is an awkward
|
|
|
|
* limitation, it should be quite straightforward to raise the maximum.
|
2007-06-13 20:45:13 +02:00
|
|
|
*/
|
|
|
|
|
2016-12-17 14:18:17 +01:00
|
|
|
/*
|
|
|
|
* Developer's notes:
|
|
|
|
*
|
2018-07-04 21:42:46 +02:00
|
|
|
* The IDA uses the functionality provided by the XArray to store bitmaps in
|
|
|
|
* each entry. The XA_FREE_MARK is only cleared when all bits in the bitmap
|
|
|
|
* have been set.
|
2016-12-17 14:18:17 +01:00
|
|
|
*
|
2018-07-04 21:42:46 +02:00
|
|
|
* I considered telling the XArray that each slot is an order-10 node
|
|
|
|
* and indexing by bit number, but the XArray can't allow a single multi-index
|
|
|
|
* entry in the head, which would significantly increase memory consumption
|
|
|
|
* for the IDA. So instead we divide the index by the number of bits in the
|
|
|
|
* leaf bitmap before doing a radix tree lookup.
|
2016-12-17 14:18:17 +01:00
|
|
|
*
|
|
|
|
* As an optimisation, if there are only a few low bits set in any given
|
2017-11-03 18:30:42 +01:00
|
|
|
* leaf, instead of allocating a 128-byte bitmap, we store the bits
|
2018-07-04 21:42:46 +02:00
|
|
|
* as a value entry. Value entries never have the XA_FREE_MARK cleared
|
|
|
|
* because we can always convert them into a bitmap entry.
|
|
|
|
*
|
|
|
|
* It would be possible to optimise further; once we've run out of a
|
|
|
|
* single 128-byte bitmap, we currently switch to a 576-byte node, put
|
|
|
|
* the 128-byte bitmap in the first entry and then start allocating extra
|
|
|
|
* 128-byte entries. We could instead use the 512 bytes of the node's
|
|
|
|
* data as a bitmap before moving to that scheme. I do not believe this
|
|
|
|
* is a worthwhile optimisation; Rasmus Villemoes surveyed the current
|
|
|
|
* users of the IDA and almost none of them use more than 1024 entries.
|
|
|
|
* Those that do use more than the 8192 IDs that the 512 bytes would
|
|
|
|
* provide.
|
|
|
|
*
|
|
|
|
* The IDA always uses a lock to alloc/free. If we add a 'test_bit'
|
2016-12-17 14:18:17 +01:00
|
|
|
* equivalent, it will still need locking. Going to RCU lookup would require
|
|
|
|
* using RCU to free bitmaps, and that's not trivial without embedding an
|
|
|
|
* RCU head in the bitmap, which adds a 2-pointer overhead to each 128-byte
|
|
|
|
* bitmap, which is excessive.
|
|
|
|
*/
|
|
|
|
|
2018-07-04 21:42:46 +02:00
|
|
|
/**
|
|
|
|
* ida_alloc_range() - Allocate an unused ID.
|
|
|
|
* @ida: IDA handle.
|
|
|
|
* @min: Lowest ID to allocate.
|
|
|
|
* @max: Highest ID to allocate.
|
|
|
|
* @gfp: Memory allocation flags.
|
|
|
|
*
|
|
|
|
* Allocate an ID between @min and @max, inclusive. The allocated ID will
|
|
|
|
* not exceed %INT_MAX, even if @max is larger.
|
|
|
|
*
|
|
|
|
* Context: Any context.
|
|
|
|
* Return: The allocated ID, or %-ENOMEM if memory could not be allocated,
|
|
|
|
* or %-ENOSPC if there are no free IDs.
|
|
|
|
*/
|
|
|
|
int ida_alloc_range(struct ida *ida, unsigned int min, unsigned int max,
|
|
|
|
gfp_t gfp)
|
2007-06-13 20:45:13 +02:00
|
|
|
{
|
2018-07-04 21:42:46 +02:00
|
|
|
XA_STATE(xas, &ida->xa, min / IDA_BITMAP_BITS);
|
|
|
|
unsigned bit = min % IDA_BITMAP_BITS;
|
|
|
|
unsigned long flags;
|
|
|
|
struct ida_bitmap *bitmap, *alloc = NULL;
|
|
|
|
|
|
|
|
if ((int)min < 0)
|
|
|
|
return -ENOSPC;
|
|
|
|
|
|
|
|
if ((int)max < 0)
|
|
|
|
max = INT_MAX;
|
|
|
|
|
|
|
|
retry:
|
|
|
|
xas_lock_irqsave(&xas, flags);
|
|
|
|
next:
|
|
|
|
bitmap = xas_find_marked(&xas, max / IDA_BITMAP_BITS, XA_FREE_MARK);
|
|
|
|
if (xas.xa_index > min / IDA_BITMAP_BITS)
|
|
|
|
bit = 0;
|
|
|
|
if (xas.xa_index * IDA_BITMAP_BITS + bit > max)
|
|
|
|
goto nospc;
|
|
|
|
|
|
|
|
if (xa_is_value(bitmap)) {
|
|
|
|
unsigned long tmp = xa_to_value(bitmap);
|
|
|
|
|
|
|
|
if (bit < BITS_PER_XA_VALUE) {
|
|
|
|
bit = find_next_zero_bit(&tmp, BITS_PER_XA_VALUE, bit);
|
|
|
|
if (xas.xa_index * IDA_BITMAP_BITS + bit > max)
|
|
|
|
goto nospc;
|
|
|
|
if (bit < BITS_PER_XA_VALUE) {
|
|
|
|
tmp |= 1UL << bit;
|
|
|
|
xas_store(&xas, xa_mk_value(tmp));
|
|
|
|
goto out;
|
2016-12-20 16:27:56 +01:00
|
|
|
}
|
|
|
|
}
|
2018-07-04 21:42:46 +02:00
|
|
|
bitmap = alloc;
|
|
|
|
if (!bitmap)
|
|
|
|
bitmap = kzalloc(sizeof(*bitmap), GFP_NOWAIT);
|
|
|
|
if (!bitmap)
|
|
|
|
goto alloc;
|
|
|
|
bitmap->bitmap[0] = tmp;
|
|
|
|
xas_store(&xas, bitmap);
|
|
|
|
if (xas_error(&xas)) {
|
|
|
|
bitmap->bitmap[0] = 0;
|
|
|
|
goto out;
|
2016-12-17 14:18:17 +01:00
|
|
|
}
|
2018-07-04 21:42:46 +02:00
|
|
|
}
|
2016-12-17 14:18:17 +01:00
|
|
|
|
2018-07-04 21:42:46 +02:00
|
|
|
if (bitmap) {
|
|
|
|
bit = find_next_zero_bit(bitmap->bitmap, IDA_BITMAP_BITS, bit);
|
|
|
|
if (xas.xa_index * IDA_BITMAP_BITS + bit > max)
|
|
|
|
goto nospc;
|
|
|
|
if (bit == IDA_BITMAP_BITS)
|
|
|
|
goto next;
|
2007-06-13 20:45:13 +02:00
|
|
|
|
2018-07-04 21:42:46 +02:00
|
|
|
__set_bit(bit, bitmap->bitmap);
|
|
|
|
if (bitmap_full(bitmap->bitmap, IDA_BITMAP_BITS))
|
|
|
|
xas_clear_mark(&xas, XA_FREE_MARK);
|
|
|
|
} else {
|
|
|
|
if (bit < BITS_PER_XA_VALUE) {
|
|
|
|
bitmap = xa_mk_value(1UL << bit);
|
2016-12-20 16:27:56 +01:00
|
|
|
} else {
|
2018-07-04 21:42:46 +02:00
|
|
|
bitmap = alloc;
|
|
|
|
if (!bitmap)
|
|
|
|
bitmap = kzalloc(sizeof(*bitmap), GFP_NOWAIT);
|
|
|
|
if (!bitmap)
|
|
|
|
goto alloc;
|
|
|
|
__set_bit(bit, bitmap->bitmap);
|
2016-12-20 16:27:56 +01:00
|
|
|
}
|
2018-07-04 21:42:46 +02:00
|
|
|
xas_store(&xas, bitmap);
|
|
|
|
}
|
|
|
|
out:
|
|
|
|
xas_unlock_irqrestore(&xas, flags);
|
|
|
|
if (xas_nomem(&xas, gfp)) {
|
|
|
|
xas.xa_index = min / IDA_BITMAP_BITS;
|
|
|
|
bit = min % IDA_BITMAP_BITS;
|
|
|
|
goto retry;
|
2007-06-13 20:45:13 +02:00
|
|
|
}
|
2018-07-04 21:42:46 +02:00
|
|
|
if (bitmap != alloc)
|
|
|
|
kfree(alloc);
|
|
|
|
if (xas_error(&xas))
|
|
|
|
return xas_error(&xas);
|
|
|
|
return xas.xa_index * IDA_BITMAP_BITS + bit;
|
|
|
|
alloc:
|
|
|
|
xas_unlock_irqrestore(&xas, flags);
|
|
|
|
alloc = kzalloc(sizeof(*bitmap), gfp);
|
|
|
|
if (!alloc)
|
|
|
|
return -ENOMEM;
|
|
|
|
xas_set(&xas, min / IDA_BITMAP_BITS);
|
|
|
|
bit = min % IDA_BITMAP_BITS;
|
|
|
|
goto retry;
|
|
|
|
nospc:
|
|
|
|
xas_unlock_irqrestore(&xas, flags);
|
|
|
|
return -ENOSPC;
|
2007-06-13 20:45:13 +02:00
|
|
|
}
|
2018-07-04 21:42:46 +02:00
|
|
|
EXPORT_SYMBOL(ida_alloc_range);
|
2007-06-13 20:45:13 +02:00
|
|
|
|
2018-07-04 21:42:46 +02:00
|
|
|
/**
|
|
|
|
* ida_free() - Release an allocated ID.
|
|
|
|
* @ida: IDA handle.
|
|
|
|
* @id: Previously allocated ID.
|
|
|
|
*
|
|
|
|
* Context: Any context.
|
|
|
|
*/
|
|
|
|
void ida_free(struct ida *ida, unsigned int id)
|
2007-06-13 20:45:13 +02:00
|
|
|
{
|
2018-07-04 21:42:46 +02:00
|
|
|
XA_STATE(xas, &ida->xa, id / IDA_BITMAP_BITS);
|
|
|
|
unsigned bit = id % IDA_BITMAP_BITS;
|
2007-06-13 20:45:13 +02:00
|
|
|
struct ida_bitmap *bitmap;
|
2018-07-04 21:42:46 +02:00
|
|
|
unsigned long flags;
|
2007-06-13 20:45:13 +02:00
|
|
|
|
2018-07-04 21:42:46 +02:00
|
|
|
BUG_ON((int)id < 0);
|
|
|
|
|
|
|
|
xas_lock_irqsave(&xas, flags);
|
|
|
|
bitmap = xas_load(&xas);
|
2014-06-06 23:37:11 +02:00
|
|
|
|
2017-11-03 18:30:42 +01:00
|
|
|
if (xa_is_value(bitmap)) {
|
2018-07-04 21:42:46 +02:00
|
|
|
unsigned long v = xa_to_value(bitmap);
|
|
|
|
if (bit >= BITS_PER_XA_VALUE)
|
|
|
|
goto err;
|
|
|
|
if (!(v & (1UL << bit)))
|
2016-12-17 14:18:17 +01:00
|
|
|
goto err;
|
2018-07-04 21:42:46 +02:00
|
|
|
v &= ~(1UL << bit);
|
|
|
|
if (!v)
|
|
|
|
goto delete;
|
|
|
|
xas_store(&xas, xa_mk_value(v));
|
2016-12-17 14:18:17 +01:00
|
|
|
} else {
|
2018-07-04 21:42:46 +02:00
|
|
|
if (!test_bit(bit, bitmap->bitmap))
|
|
|
|
goto err;
|
|
|
|
__clear_bit(bit, bitmap->bitmap);
|
|
|
|
xas_set_mark(&xas, XA_FREE_MARK);
|
|
|
|
if (bitmap_empty(bitmap->bitmap, IDA_BITMAP_BITS)) {
|
|
|
|
kfree(bitmap);
|
|
|
|
delete:
|
|
|
|
xas_store(&xas, NULL);
|
|
|
|
}
|
2007-06-13 20:45:13 +02:00
|
|
|
}
|
2018-07-04 21:42:46 +02:00
|
|
|
xas_unlock_irqrestore(&xas, flags);
|
2007-06-13 20:45:13 +02:00
|
|
|
return;
|
|
|
|
err:
|
2018-07-04 21:42:46 +02:00
|
|
|
xas_unlock_irqrestore(&xas, flags);
|
2018-06-19 01:02:48 +02:00
|
|
|
WARN(1, "ida_free called for id=%d which is not allocated.\n", id);
|
2007-06-13 20:45:13 +02:00
|
|
|
}
|
2018-07-04 21:42:46 +02:00
|
|
|
EXPORT_SYMBOL(ida_free);
|
2007-06-13 20:45:13 +02:00
|
|
|
|
|
|
|
/**
|
2018-06-21 21:36:45 +02:00
|
|
|
* ida_destroy() - Free all IDs.
|
|
|
|
* @ida: IDA handle.
|
|
|
|
*
|
|
|
|
* Calling this function frees all IDs and releases all resources used
|
|
|
|
* by an IDA. When this call returns, the IDA is empty and can be reused
|
|
|
|
* or freed. If the IDA is already empty, there is no need to call this
|
|
|
|
* function.
|
2016-12-20 16:27:56 +01:00
|
|
|
*
|
2018-06-21 21:36:45 +02:00
|
|
|
* Context: Any context.
|
2007-06-13 20:45:13 +02:00
|
|
|
*/
|
|
|
|
void ida_destroy(struct ida *ida)
|
|
|
|
{
|
2018-07-04 21:42:46 +02:00
|
|
|
XA_STATE(xas, &ida->xa, 0);
|
|
|
|
struct ida_bitmap *bitmap;
|
2018-06-21 21:36:45 +02:00
|
|
|
unsigned long flags;
|
2016-12-20 16:27:56 +01:00
|
|
|
|
2018-07-04 21:42:46 +02:00
|
|
|
xas_lock_irqsave(&xas, flags);
|
|
|
|
xas_for_each(&xas, bitmap, ULONG_MAX) {
|
2017-11-03 18:30:42 +01:00
|
|
|
if (!xa_is_value(bitmap))
|
2016-12-17 14:18:17 +01:00
|
|
|
kfree(bitmap);
|
2018-07-04 21:42:46 +02:00
|
|
|
xas_store(&xas, NULL);
|
2016-12-20 16:27:56 +01:00
|
|
|
}
|
2018-07-04 21:42:46 +02:00
|
|
|
xas_unlock_irqrestore(&xas, flags);
|
2007-06-13 20:45:13 +02:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(ida_destroy);
|
|
|
|
|
2018-07-04 21:42:46 +02:00
|
|
|
#ifndef __KERNEL__
|
|
|
|
extern void xa_dump_index(unsigned long index, unsigned int shift);
|
|
|
|
#define IDA_CHUNK_SHIFT ilog2(IDA_BITMAP_BITS)
|
2011-08-04 01:21:06 +02:00
|
|
|
|
2018-07-04 21:42:46 +02:00
|
|
|
static void ida_dump_entry(void *entry, unsigned long index)
|
|
|
|
{
|
|
|
|
unsigned long i;
|
|
|
|
|
|
|
|
if (!entry)
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (xa_is_node(entry)) {
|
|
|
|
struct xa_node *node = xa_to_node(entry);
|
|
|
|
unsigned int shift = node->shift + IDA_CHUNK_SHIFT +
|
|
|
|
XA_CHUNK_SHIFT;
|
|
|
|
|
|
|
|
xa_dump_index(index * IDA_BITMAP_BITS, shift);
|
|
|
|
xa_dump_node(node);
|
|
|
|
for (i = 0; i < XA_CHUNK_SIZE; i++)
|
|
|
|
ida_dump_entry(node->slots[i],
|
|
|
|
index | (i << node->shift));
|
|
|
|
} else if (xa_is_value(entry)) {
|
|
|
|
xa_dump_index(index * IDA_BITMAP_BITS, ilog2(BITS_PER_LONG));
|
|
|
|
pr_cont("value: data %lx [%px]\n", xa_to_value(entry), entry);
|
|
|
|
} else {
|
|
|
|
struct ida_bitmap *bitmap = entry;
|
2011-08-04 01:21:06 +02:00
|
|
|
|
2018-07-04 21:42:46 +02:00
|
|
|
xa_dump_index(index * IDA_BITMAP_BITS, IDA_CHUNK_SHIFT);
|
|
|
|
pr_cont("bitmap: %p data", bitmap);
|
|
|
|
for (i = 0; i < IDA_BITMAP_LONGS; i++)
|
|
|
|
pr_cont(" %lx", bitmap->bitmap[i]);
|
|
|
|
pr_cont("\n");
|
2018-03-20 22:07:11 +01:00
|
|
|
}
|
2011-08-04 01:21:06 +02:00
|
|
|
}
|
|
|
|
|
2018-07-04 21:42:46 +02:00
|
|
|
static void ida_dump(struct ida *ida)
|
2011-08-04 01:21:06 +02:00
|
|
|
{
|
2018-07-04 21:42:46 +02:00
|
|
|
struct xarray *xa = &ida->xa;
|
|
|
|
pr_debug("ida: %p node %p free %d\n", ida, xa->xa_head,
|
|
|
|
xa->xa_flags >> ROOT_TAG_SHIFT);
|
|
|
|
ida_dump_entry(xa->xa_head, 0);
|
2011-08-04 01:21:06 +02:00
|
|
|
}
|
2018-07-04 21:42:46 +02:00
|
|
|
#endif
|