lglocks-rt.patch
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
This commit is contained in:
parent
22775b1bef
commit
121435cad6
|
@ -32,22 +32,39 @@
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
struct lglock {
|
struct lglock {
|
||||||
|
#ifndef CONFIG_PREEMPT_RT_FULL
|
||||||
arch_spinlock_t __percpu *lock;
|
arch_spinlock_t __percpu *lock;
|
||||||
|
#else
|
||||||
|
struct rt_mutex __percpu *lock;
|
||||||
|
#endif
|
||||||
#ifdef CONFIG_DEBUG_LOCK_ALLOC
|
#ifdef CONFIG_DEBUG_LOCK_ALLOC
|
||||||
struct lock_class_key lock_key;
|
struct lock_class_key lock_key;
|
||||||
struct lockdep_map lock_dep_map;
|
struct lockdep_map lock_dep_map;
|
||||||
#endif
|
#endif
|
||||||
};
|
};
|
||||||
|
|
||||||
#define DEFINE_LGLOCK(name) \
|
#ifndef CONFIG_PREEMPT_RT_FULL
|
||||||
|
# define DEFINE_LGLOCK(name) \
|
||||||
static DEFINE_PER_CPU(arch_spinlock_t, name ## _lock) \
|
static DEFINE_PER_CPU(arch_spinlock_t, name ## _lock) \
|
||||||
= __ARCH_SPIN_LOCK_UNLOCKED; \
|
= __ARCH_SPIN_LOCK_UNLOCKED; \
|
||||||
struct lglock name = { .lock = &name ## _lock }
|
struct lglock name = { .lock = &name ## _lock }
|
||||||
|
|
||||||
#define DEFINE_STATIC_LGLOCK(name) \
|
# define DEFINE_STATIC_LGLOCK(name) \
|
||||||
static DEFINE_PER_CPU(arch_spinlock_t, name ## _lock) \
|
static DEFINE_PER_CPU(arch_spinlock_t, name ## _lock) \
|
||||||
= __ARCH_SPIN_LOCK_UNLOCKED; \
|
= __ARCH_SPIN_LOCK_UNLOCKED; \
|
||||||
static struct lglock name = { .lock = &name ## _lock }
|
static struct lglock name = { .lock = &name ## _lock }
|
||||||
|
#else
|
||||||
|
|
||||||
|
# define DEFINE_LGLOCK(name) \
|
||||||
|
static DEFINE_PER_CPU(struct rt_mutex, name ## _lock) \
|
||||||
|
= __RT_MUTEX_INITIALIZER( name ## _lock); \
|
||||||
|
struct lglock name = { .lock = &name ## _lock }
|
||||||
|
|
||||||
|
# define DEFINE_STATIC_LGLOCK(name) \
|
||||||
|
static DEFINE_PER_CPU(struct rt_mutex, name ## _lock) \
|
||||||
|
= __RT_MUTEX_INITIALIZER( name ## _lock); \
|
||||||
|
static struct lglock name = { .lock = &name ## _lock }
|
||||||
|
#endif
|
||||||
|
|
||||||
void lg_lock_init(struct lglock *lg, char *name);
|
void lg_lock_init(struct lglock *lg, char *name);
|
||||||
void lg_local_lock(struct lglock *lg);
|
void lg_local_lock(struct lglock *lg);
|
||||||
|
|
|
@ -4,6 +4,15 @@
|
||||||
#include <linux/cpu.h>
|
#include <linux/cpu.h>
|
||||||
#include <linux/string.h>
|
#include <linux/string.h>
|
||||||
|
|
||||||
|
#ifndef CONFIG_PREEMPT_RT_FULL
|
||||||
|
# define lg_lock_ptr arch_spinlock_t
|
||||||
|
# define lg_do_lock(l) arch_spin_lock(l)
|
||||||
|
# define lg_do_unlock(l) arch_spin_unlock(l)
|
||||||
|
#else
|
||||||
|
# define lg_lock_ptr struct rt_mutex
|
||||||
|
# define lg_do_lock(l) __rt_spin_lock(l)
|
||||||
|
# define lg_do_unlock(l) __rt_spin_unlock(l)
|
||||||
|
#endif
|
||||||
/*
|
/*
|
||||||
* Note there is no uninit, so lglocks cannot be defined in
|
* Note there is no uninit, so lglocks cannot be defined in
|
||||||
* modules (but it's fine to use them from there)
|
* modules (but it's fine to use them from there)
|
||||||
|
@ -12,51 +21,60 @@
|
||||||
|
|
||||||
void lg_lock_init(struct lglock *lg, char *name)
|
void lg_lock_init(struct lglock *lg, char *name)
|
||||||
{
|
{
|
||||||
|
#ifdef CONFIG_PREEMPT_RT_FULL
|
||||||
|
int i;
|
||||||
|
|
||||||
|
for_each_possible_cpu(i) {
|
||||||
|
struct rt_mutex *lock = per_cpu_ptr(lg->lock, i);
|
||||||
|
|
||||||
|
rt_mutex_init(lock);
|
||||||
|
}
|
||||||
|
#endif
|
||||||
LOCKDEP_INIT_MAP(&lg->lock_dep_map, name, &lg->lock_key, 0);
|
LOCKDEP_INIT_MAP(&lg->lock_dep_map, name, &lg->lock_key, 0);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(lg_lock_init);
|
EXPORT_SYMBOL(lg_lock_init);
|
||||||
|
|
||||||
void lg_local_lock(struct lglock *lg)
|
void lg_local_lock(struct lglock *lg)
|
||||||
{
|
{
|
||||||
arch_spinlock_t *lock;
|
lg_lock_ptr *lock;
|
||||||
|
|
||||||
preempt_disable();
|
migrate_disable();
|
||||||
lock_acquire_shared(&lg->lock_dep_map, 0, 0, NULL, _RET_IP_);
|
lock_acquire_shared(&lg->lock_dep_map, 0, 0, NULL, _RET_IP_);
|
||||||
lock = this_cpu_ptr(lg->lock);
|
lock = this_cpu_ptr(lg->lock);
|
||||||
arch_spin_lock(lock);
|
lg_do_lock(lock);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(lg_local_lock);
|
EXPORT_SYMBOL(lg_local_lock);
|
||||||
|
|
||||||
void lg_local_unlock(struct lglock *lg)
|
void lg_local_unlock(struct lglock *lg)
|
||||||
{
|
{
|
||||||
arch_spinlock_t *lock;
|
lg_lock_ptr *lock;
|
||||||
|
|
||||||
lock_release(&lg->lock_dep_map, 1, _RET_IP_);
|
lock_release(&lg->lock_dep_map, 1, _RET_IP_);
|
||||||
lock = this_cpu_ptr(lg->lock);
|
lock = this_cpu_ptr(lg->lock);
|
||||||
arch_spin_unlock(lock);
|
lg_do_unlock(lock);
|
||||||
preempt_enable();
|
migrate_enable();
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(lg_local_unlock);
|
EXPORT_SYMBOL(lg_local_unlock);
|
||||||
|
|
||||||
void lg_local_lock_cpu(struct lglock *lg, int cpu)
|
void lg_local_lock_cpu(struct lglock *lg, int cpu)
|
||||||
{
|
{
|
||||||
arch_spinlock_t *lock;
|
lg_lock_ptr *lock;
|
||||||
|
|
||||||
preempt_disable();
|
preempt_disable_nort();
|
||||||
lock_acquire_shared(&lg->lock_dep_map, 0, 0, NULL, _RET_IP_);
|
lock_acquire_shared(&lg->lock_dep_map, 0, 0, NULL, _RET_IP_);
|
||||||
lock = per_cpu_ptr(lg->lock, cpu);
|
lock = per_cpu_ptr(lg->lock, cpu);
|
||||||
arch_spin_lock(lock);
|
lg_do_lock(lock);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(lg_local_lock_cpu);
|
EXPORT_SYMBOL(lg_local_lock_cpu);
|
||||||
|
|
||||||
void lg_local_unlock_cpu(struct lglock *lg, int cpu)
|
void lg_local_unlock_cpu(struct lglock *lg, int cpu)
|
||||||
{
|
{
|
||||||
arch_spinlock_t *lock;
|
lg_lock_ptr *lock;
|
||||||
|
|
||||||
lock_release(&lg->lock_dep_map, 1, _RET_IP_);
|
lock_release(&lg->lock_dep_map, 1, _RET_IP_);
|
||||||
lock = per_cpu_ptr(lg->lock, cpu);
|
lock = per_cpu_ptr(lg->lock, cpu);
|
||||||
arch_spin_unlock(lock);
|
lg_do_unlock(lock);
|
||||||
preempt_enable();
|
preempt_enable_nort();
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(lg_local_unlock_cpu);
|
EXPORT_SYMBOL(lg_local_unlock_cpu);
|
||||||
|
|
||||||
|
@ -64,12 +82,12 @@ void lg_global_lock(struct lglock *lg)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
preempt_disable();
|
preempt_disable_nort();
|
||||||
lock_acquire_exclusive(&lg->lock_dep_map, 0, 0, NULL, _RET_IP_);
|
lock_acquire_exclusive(&lg->lock_dep_map, 0, 0, NULL, _RET_IP_);
|
||||||
for_each_possible_cpu(i) {
|
for_each_possible_cpu(i) {
|
||||||
arch_spinlock_t *lock;
|
lg_lock_ptr *lock;
|
||||||
lock = per_cpu_ptr(lg->lock, i);
|
lock = per_cpu_ptr(lg->lock, i);
|
||||||
arch_spin_lock(lock);
|
lg_do_lock(lock);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(lg_global_lock);
|
EXPORT_SYMBOL(lg_global_lock);
|
||||||
|
@ -80,10 +98,10 @@ void lg_global_unlock(struct lglock *lg)
|
||||||
|
|
||||||
lock_release(&lg->lock_dep_map, 1, _RET_IP_);
|
lock_release(&lg->lock_dep_map, 1, _RET_IP_);
|
||||||
for_each_possible_cpu(i) {
|
for_each_possible_cpu(i) {
|
||||||
arch_spinlock_t *lock;
|
lg_lock_ptr *lock;
|
||||||
lock = per_cpu_ptr(lg->lock, i);
|
lock = per_cpu_ptr(lg->lock, i);
|
||||||
arch_spin_unlock(lock);
|
lg_do_unlock(lock);
|
||||||
}
|
}
|
||||||
preempt_enable();
|
preempt_enable_nort();
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(lg_global_unlock);
|
EXPORT_SYMBOL(lg_global_unlock);
|
||||||
|
|
Loading…
Reference in New Issue