提交 bd197234 编写于 作者: T Thomas Gleixner 提交者: Linus Torvalds

Revert "futex_requeue_pi optimization"

This reverts commit d0aa7a70.

It not only introduced user space visible changes to the futex syscall,
it is also non-functional and there is no way to fix it proper before
the 2.6.22 release.

The breakage report ( http://lkml.org/lkml/2007/5/12/17 ) went
unanswered, and unfortunately it turned out that the concept is not
feasible at all.  It violates the rtmutex semantics badly by introducing
a virtual owner, which hacks around the coupling of the user-space
pi_futex and the kernel internal rt_mutex representation.

At the moment the only safe option is to remove it fully as it contains
user-space visible changes to broken kernel code, which we do not want
to expose in the 2.6.22 release.

The patch reverts the original patch mostly 1:1, but contains a couple
of trivial manual cleanups which were necessary due to patches, which
touched the same area of code later.

Verified against the glibc tests and my own PI futex tests.
Signed-off-by: NThomas Gleixner <tglx@linutronix.de>
Acked-by: NIngo Molnar <mingo@elte.hu>
Acked-by: NUlrich Drepper <drepper@redhat.com>
Cc: Pierre Peiffer <pierre.peiffer@bull.net>
Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
上级 188e1f81
...@@ -17,7 +17,6 @@ union ktime; ...@@ -17,7 +17,6 @@ union ktime;
#define FUTEX_LOCK_PI 6 #define FUTEX_LOCK_PI 6
#define FUTEX_UNLOCK_PI 7 #define FUTEX_UNLOCK_PI 7
#define FUTEX_TRYLOCK_PI 8 #define FUTEX_TRYLOCK_PI 8
#define FUTEX_CMP_REQUEUE_PI 9
#define FUTEX_PRIVATE_FLAG 128 #define FUTEX_PRIVATE_FLAG 128
#define FUTEX_CMD_MASK ~FUTEX_PRIVATE_FLAG #define FUTEX_CMD_MASK ~FUTEX_PRIVATE_FLAG
...@@ -97,15 +96,10 @@ struct robust_list_head { ...@@ -97,15 +96,10 @@ struct robust_list_head {
*/ */
#define FUTEX_OWNER_DIED 0x40000000 #define FUTEX_OWNER_DIED 0x40000000
/*
* Some processes have been requeued on this PI-futex
*/
#define FUTEX_WAITER_REQUEUED 0x20000000
/* /*
* The rest of the robust-futex field is for the TID: * The rest of the robust-futex field is for the TID:
*/ */
#define FUTEX_TID_MASK 0x0fffffff #define FUTEX_TID_MASK 0x3fffffff
/* /*
* This limit protects against a deliberately circular list. * This limit protects against a deliberately circular list.
...@@ -139,7 +133,6 @@ handle_futex_death(u32 __user *uaddr, struct task_struct *curr, int pi); ...@@ -139,7 +133,6 @@ handle_futex_death(u32 __user *uaddr, struct task_struct *curr, int pi);
#define FUT_OFF_MMSHARED 2 /* We set bit 1 if key has a reference on mm */ #define FUT_OFF_MMSHARED 2 /* We set bit 1 if key has a reference on mm */
union futex_key { union futex_key {
u32 __user *uaddr;
struct { struct {
unsigned long pgoff; unsigned long pgoff;
struct inode *inode; struct inode *inode;
......
...@@ -56,12 +56,6 @@ ...@@ -56,12 +56,6 @@
#include "rtmutex_common.h" #include "rtmutex_common.h"
#ifdef CONFIG_DEBUG_RT_MUTEXES
# include "rtmutex-debug.h"
#else
# include "rtmutex.h"
#endif
#define FUTEX_HASHBITS (CONFIG_BASE_SMALL ? 4 : 8) #define FUTEX_HASHBITS (CONFIG_BASE_SMALL ? 4 : 8)
/* /*
...@@ -111,12 +105,6 @@ struct futex_q { ...@@ -111,12 +105,6 @@ struct futex_q {
/* Optional priority inheritance state: */ /* Optional priority inheritance state: */
struct futex_pi_state *pi_state; struct futex_pi_state *pi_state;
struct task_struct *task; struct task_struct *task;
/*
* This waiter is used in case of requeue from a
* normal futex to a PI-futex
*/
struct rt_mutex_waiter waiter;
}; };
/* /*
...@@ -216,9 +204,6 @@ int get_futex_key(u32 __user *uaddr, struct rw_semaphore *fshared, ...@@ -216,9 +204,6 @@ int get_futex_key(u32 __user *uaddr, struct rw_semaphore *fshared,
if (unlikely((vma->vm_flags & (VM_IO|VM_READ)) != VM_READ)) if (unlikely((vma->vm_flags & (VM_IO|VM_READ)) != VM_READ))
return (vma->vm_flags & VM_IO) ? -EPERM : -EACCES; return (vma->vm_flags & VM_IO) ? -EPERM : -EACCES;
/* Save the user address in the ley */
key->uaddr = uaddr;
/* /*
* Private mappings are handled in a simple way. * Private mappings are handled in a simple way.
* *
...@@ -636,8 +621,6 @@ static int wake_futex_pi(u32 __user *uaddr, u32 uval, struct futex_q *this) ...@@ -636,8 +621,6 @@ static int wake_futex_pi(u32 __user *uaddr, u32 uval, struct futex_q *this)
int ret = 0; int ret = 0;
newval = FUTEX_WAITERS | new_owner->pid; newval = FUTEX_WAITERS | new_owner->pid;
/* Keep the FUTEX_WAITER_REQUEUED flag if it was set */
newval |= (uval & FUTEX_WAITER_REQUEUED);
pagefault_disable(); pagefault_disable();
curval = futex_atomic_cmpxchg_inatomic(uaddr, uval, newval); curval = futex_atomic_cmpxchg_inatomic(uaddr, uval, newval);
...@@ -749,259 +732,6 @@ static int futex_wake(u32 __user *uaddr, struct rw_semaphore *fshared, ...@@ -749,259 +732,6 @@ static int futex_wake(u32 __user *uaddr, struct rw_semaphore *fshared,
return ret; return ret;
} }
/*
* Called from futex_requeue_pi.
* Set FUTEX_WAITERS and FUTEX_WAITER_REQUEUED flags on the
* PI-futex value; search its associated pi_state if an owner exist
* or create a new one without owner.
*/
static inline int
lookup_pi_state_for_requeue(u32 __user *uaddr, struct futex_hash_bucket *hb,
union futex_key *key,
struct futex_pi_state **pi_state)
{
u32 curval, uval, newval;
retry:
/*
* We can't handle a fault cleanly because we can't
* release the locks here. Simply return the fault.
*/
if (get_futex_value_locked(&curval, uaddr))
return -EFAULT;
/* set the flags FUTEX_WAITERS and FUTEX_WAITER_REQUEUED */
if ((curval & (FUTEX_WAITERS | FUTEX_WAITER_REQUEUED))
!= (FUTEX_WAITERS | FUTEX_WAITER_REQUEUED)) {
/*
* No waiters yet, we prepare the futex to have some waiters.
*/
uval = curval;
newval = uval | FUTEX_WAITERS | FUTEX_WAITER_REQUEUED;
pagefault_disable();
curval = futex_atomic_cmpxchg_inatomic(uaddr, uval, newval);
pagefault_enable();
if (unlikely(curval == -EFAULT))
return -EFAULT;
if (unlikely(curval != uval))
goto retry;
}
if (!(curval & FUTEX_TID_MASK)
|| lookup_pi_state(curval, hb, key, pi_state)) {
/* the futex has no owner (yet) or the lookup failed:
allocate one pi_state without owner */
*pi_state = alloc_pi_state();
/* Already stores the key: */
(*pi_state)->key = *key;
/* init the mutex without owner */
__rt_mutex_init(&(*pi_state)->pi_mutex, NULL);
}
return 0;
}
/*
* Keep the first nr_wake waiter from futex1, wake up one,
* and requeue the next nr_requeue waiters following hashed on
* one physical page to another physical page (PI-futex uaddr2)
*/
static int futex_requeue_pi(u32 __user *uaddr1,
struct rw_semaphore *fshared,
u32 __user *uaddr2,
int nr_wake, int nr_requeue, u32 *cmpval)
{
union futex_key key1, key2;
struct futex_hash_bucket *hb1, *hb2;
struct plist_head *head1;
struct futex_q *this, *next;
struct futex_pi_state *pi_state2 = NULL;
struct rt_mutex_waiter *waiter, *top_waiter = NULL;
struct rt_mutex *lock2 = NULL;
int ret, drop_count = 0;
if (refill_pi_state_cache())
return -ENOMEM;
retry:
/*
* First take all the futex related locks:
*/
if (fshared)
down_read(fshared);
ret = get_futex_key(uaddr1, fshared, &key1);
if (unlikely(ret != 0))
goto out;
ret = get_futex_key(uaddr2, fshared, &key2);
if (unlikely(ret != 0))
goto out;
hb1 = hash_futex(&key1);
hb2 = hash_futex(&key2);
double_lock_hb(hb1, hb2);
if (likely(cmpval != NULL)) {
u32 curval;
ret = get_futex_value_locked(&curval, uaddr1);
if (unlikely(ret)) {
spin_unlock(&hb1->lock);
if (hb1 != hb2)
spin_unlock(&hb2->lock);
/*
* If we would have faulted, release mmap_sem, fault
* it in and start all over again.
*/
if (fshared)
up_read(fshared);
ret = get_user(curval, uaddr1);
if (!ret)
goto retry;
return ret;
}
if (curval != *cmpval) {
ret = -EAGAIN;
goto out_unlock;
}
}
head1 = &hb1->chain;
plist_for_each_entry_safe(this, next, head1, list) {
if (!match_futex (&this->key, &key1))
continue;
if (++ret <= nr_wake) {
wake_futex(this);
} else {
/*
* FIRST: get and set the pi_state
*/
if (!pi_state2) {
int s;
/* do this only the first time we requeue someone */
s = lookup_pi_state_for_requeue(uaddr2, hb2,
&key2, &pi_state2);
if (s) {
ret = s;
goto out_unlock;
}
lock2 = &pi_state2->pi_mutex;
spin_lock(&lock2->wait_lock);
/* Save the top waiter of the wait_list */
if (rt_mutex_has_waiters(lock2))
top_waiter = rt_mutex_top_waiter(lock2);
} else
atomic_inc(&pi_state2->refcount);
this->pi_state = pi_state2;
/*
* SECOND: requeue futex_q to the correct hashbucket
*/
/*
* If key1 and key2 hash to the same bucket, no need to
* requeue.
*/
if (likely(head1 != &hb2->chain)) {
plist_del(&this->list, &hb1->chain);
plist_add(&this->list, &hb2->chain);
this->lock_ptr = &hb2->lock;
#ifdef CONFIG_DEBUG_PI_LIST
this->list.plist.lock = &hb2->lock;
#endif
}
this->key = key2;
get_futex_key_refs(&key2);
drop_count++;
/*
* THIRD: queue it to lock2
*/
spin_lock_irq(&this->task->pi_lock);
waiter = &this->waiter;
waiter->task = this->task;
waiter->lock = lock2;
plist_node_init(&waiter->list_entry, this->task->prio);
plist_node_init(&waiter->pi_list_entry, this->task->prio);
plist_add(&waiter->list_entry, &lock2->wait_list);
this->task->pi_blocked_on = waiter;
spin_unlock_irq(&this->task->pi_lock);
if (ret - nr_wake >= nr_requeue)
break;
}
}
/* If we've requeued some tasks and the top_waiter of the rt_mutex
has changed, we must adjust the priority of the owner, if any */
if (drop_count) {
struct task_struct *owner = rt_mutex_owner(lock2);
if (owner &&
(top_waiter != (waiter = rt_mutex_top_waiter(lock2)))) {
int chain_walk = 0;
spin_lock_irq(&owner->pi_lock);
if (top_waiter)
plist_del(&top_waiter->pi_list_entry, &owner->pi_waiters);
else
/*
* There was no waiters before the requeue,
* the flag must be updated
*/
mark_rt_mutex_waiters(lock2);
plist_add(&waiter->pi_list_entry, &owner->pi_waiters);
__rt_mutex_adjust_prio(owner);
if (owner->pi_blocked_on) {
chain_walk = 1;
get_task_struct(owner);
}
spin_unlock_irq(&owner->pi_lock);
spin_unlock(&lock2->wait_lock);
if (chain_walk)
rt_mutex_adjust_prio_chain(owner, 0, lock2, NULL,
current);
} else {
/* No owner or the top_waiter does not change */
mark_rt_mutex_waiters(lock2);
spin_unlock(&lock2->wait_lock);
}
}
out_unlock:
spin_unlock(&hb1->lock);
if (hb1 != hb2)
spin_unlock(&hb2->lock);
/* drop_futex_key_refs() must be called outside the spinlocks. */
while (--drop_count >= 0)
drop_futex_key_refs(&key1);
out:
if (fshared)
up_read(fshared);
return ret;
}
/* /*
* Wake up all waiters hashed on the physical page that is mapped * Wake up all waiters hashed on the physical page that is mapped
* to this virtual address: * to this virtual address:
...@@ -1384,7 +1114,6 @@ static int fixup_pi_state_owner(u32 __user *uaddr, struct futex_q *q, ...@@ -1384,7 +1114,6 @@ static int fixup_pi_state_owner(u32 __user *uaddr, struct futex_q *q,
while (!ret) { while (!ret) {
newval = (uval & FUTEX_OWNER_DIED) | newtid; newval = (uval & FUTEX_OWNER_DIED) | newtid;
newval |= (uval & FUTEX_WAITER_REQUEUED);
pagefault_disable(); pagefault_disable();
curval = futex_atomic_cmpxchg_inatomic(uaddr, curval = futex_atomic_cmpxchg_inatomic(uaddr,
...@@ -1416,7 +1145,7 @@ static int futex_wait(u32 __user *uaddr, struct rw_semaphore *fshared, ...@@ -1416,7 +1145,7 @@ static int futex_wait(u32 __user *uaddr, struct rw_semaphore *fshared,
struct futex_q q; struct futex_q q;
u32 uval; u32 uval;
int ret; int ret;
struct hrtimer_sleeper t, *to = NULL; struct hrtimer_sleeper t;
int rem = 0; int rem = 0;
q.pi_state = NULL; q.pi_state = NULL;
...@@ -1472,14 +1201,6 @@ static int futex_wait(u32 __user *uaddr, struct rw_semaphore *fshared, ...@@ -1472,14 +1201,6 @@ static int futex_wait(u32 __user *uaddr, struct rw_semaphore *fshared,
if (uval != val) if (uval != val)
goto out_unlock_release_sem; goto out_unlock_release_sem;
/*
* This rt_mutex_waiter structure is prepared here and will
* be used only if this task is requeued from a normal futex to
* a PI-futex with futex_requeue_pi.
*/
debug_rt_mutex_init_waiter(&q.waiter);
q.waiter.task = NULL;
/* Only actually queue if *uaddr contained val. */ /* Only actually queue if *uaddr contained val. */
__queue_me(&q, hb); __queue_me(&q, hb);
...@@ -1510,7 +1231,6 @@ static int futex_wait(u32 __user *uaddr, struct rw_semaphore *fshared, ...@@ -1510,7 +1231,6 @@ static int futex_wait(u32 __user *uaddr, struct rw_semaphore *fshared,
if (!abs_time) if (!abs_time)
schedule(); schedule();
else { else {
to = &t;
hrtimer_init(&t.timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS); hrtimer_init(&t.timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
hrtimer_init_sleeper(&t, current); hrtimer_init_sleeper(&t, current);
t.timer.expires = *abs_time; t.timer.expires = *abs_time;
...@@ -1538,67 +1258,6 @@ static int futex_wait(u32 __user *uaddr, struct rw_semaphore *fshared, ...@@ -1538,67 +1258,6 @@ static int futex_wait(u32 __user *uaddr, struct rw_semaphore *fshared,
* we are the only user of it. * we are the only user of it.
*/ */
if (q.pi_state) {
/*
* We were woken but have been requeued on a PI-futex.
* We have to complete the lock acquisition by taking
* the rtmutex.
*/
struct rt_mutex *lock = &q.pi_state->pi_mutex;
spin_lock(&lock->wait_lock);
if (unlikely(q.waiter.task)) {
remove_waiter(lock, &q.waiter);
}
spin_unlock(&lock->wait_lock);
if (rem)
ret = -ETIMEDOUT;
else
ret = rt_mutex_timed_lock(lock, to, 1);
if (fshared)
down_read(fshared);
spin_lock(q.lock_ptr);
/*
* Got the lock. We might not be the anticipated owner if we
* did a lock-steal - fix up the PI-state in that case.
*/
if (!ret && q.pi_state->owner != curr) {
/*
* We MUST play with the futex we were requeued on,
* NOT the current futex.
* We can retrieve it from the key of the pi_state
*/
uaddr = q.pi_state->key.uaddr;
ret = fixup_pi_state_owner(uaddr, &q, curr);
} else {
/*
* Catch the rare case, where the lock was released
* when we were on the way back before we locked
* the hash bucket.
*/
if (ret && q.pi_state->owner == curr) {
if (rt_mutex_trylock(&q.pi_state->pi_mutex))
ret = 0;
}
}
/* Unqueue and drop the lock */
unqueue_me_pi(&q);
if (fshared)
up_read(fshared);
debug_rt_mutex_free_waiter(&q.waiter);
return ret;
}
debug_rt_mutex_free_waiter(&q.waiter);
/* If we were woken (and unqueued), we succeeded, whatever. */ /* If we were woken (and unqueued), we succeeded, whatever. */
if (!unqueue_me(&q)) if (!unqueue_me(&q))
return 0; return 0;
...@@ -1648,51 +1307,6 @@ static long futex_wait_restart(struct restart_block *restart) ...@@ -1648,51 +1307,6 @@ static long futex_wait_restart(struct restart_block *restart)
} }
static void set_pi_futex_owner(struct futex_hash_bucket *hb,
union futex_key *key, struct task_struct *p)
{
struct plist_head *head;
struct futex_q *this, *next;
struct futex_pi_state *pi_state = NULL;
struct rt_mutex *lock;
/* Search a waiter that should already exists */
head = &hb->chain;
plist_for_each_entry_safe(this, next, head, list) {
if (match_futex (&this->key, key)) {
pi_state = this->pi_state;
break;
}
}
BUG_ON(!pi_state);
/* set p as pi_state's owner */
lock = &pi_state->pi_mutex;
spin_lock(&lock->wait_lock);
spin_lock_irq(&p->pi_lock);
list_add(&pi_state->list, &p->pi_state_list);
pi_state->owner = p;
/* set p as pi_mutex's owner */
debug_rt_mutex_proxy_lock(lock, p);
WARN_ON(rt_mutex_owner(lock));
rt_mutex_set_owner(lock, p, 0);
rt_mutex_deadlock_account_lock(lock, p);
plist_add(&rt_mutex_top_waiter(lock)->pi_list_entry,
&p->pi_waiters);
__rt_mutex_adjust_prio(p);
spin_unlock_irq(&p->pi_lock);
spin_unlock(&lock->wait_lock);
}
/* /*
* Userspace tried a 0 -> TID atomic transition of the futex value * Userspace tried a 0 -> TID atomic transition of the futex value
* and failed. The kernel side here does the whole locking operation: * and failed. The kernel side here does the whole locking operation:
...@@ -1753,8 +1367,7 @@ static int futex_lock_pi(u32 __user *uaddr, struct rw_semaphore *fshared, ...@@ -1753,8 +1367,7 @@ static int futex_lock_pi(u32 __user *uaddr, struct rw_semaphore *fshared,
* situation and we return success to user space. * situation and we return success to user space.
*/ */
if (unlikely((curval & FUTEX_TID_MASK) == current->pid)) { if (unlikely((curval & FUTEX_TID_MASK) == current->pid)) {
if (!(curval & FUTEX_WAITER_REQUEUED)) ret = -EDEADLK;
ret = -EDEADLK;
goto out_unlock_release_sem; goto out_unlock_release_sem;
} }
...@@ -1774,14 +1387,14 @@ static int futex_lock_pi(u32 __user *uaddr, struct rw_semaphore *fshared, ...@@ -1774,14 +1387,14 @@ static int futex_lock_pi(u32 __user *uaddr, struct rw_semaphore *fshared,
/* /*
* There are two cases, where a futex might have no owner (the * There are two cases, where a futex might have no owner (the
* owner TID is 0): OWNER_DIED or REQUEUE. We take over the * owner TID is 0): OWNER_DIED. We take over the futex in this
* futex in this case. We also do an unconditional take over, * case. We also do an unconditional take over, when the owner
* when the owner of the futex died. * of the futex died.
* *
* This is safe as we are protected by the hash bucket lock ! * This is safe as we are protected by the hash bucket lock !
*/ */
if (unlikely(ownerdied || !(curval & FUTEX_TID_MASK))) { if (unlikely(ownerdied || !(curval & FUTEX_TID_MASK))) {
/* Keep the OWNER_DIED and REQUEUE bits */ /* Keep the OWNER_DIED bit */
newval = (curval & ~FUTEX_TID_MASK) | current->pid; newval = (curval & ~FUTEX_TID_MASK) | current->pid;
ownerdied = 0; ownerdied = 0;
lock_taken = 1; lock_taken = 1;
...@@ -1797,14 +1410,10 @@ static int futex_lock_pi(u32 __user *uaddr, struct rw_semaphore *fshared, ...@@ -1797,14 +1410,10 @@ static int futex_lock_pi(u32 __user *uaddr, struct rw_semaphore *fshared,
goto retry_locked; goto retry_locked;
/* /*
* We took the lock due to requeue or owner died take over. * We took the lock due to owner died take over.
*/ */
if (unlikely(lock_taken)) { if (unlikely(lock_taken))
/* For requeue we need to fixup the pi_futex */
if (curval & FUTEX_WAITER_REQUEUED)
set_pi_futex_owner(hb, &q.key, curr);
goto out_unlock_release_sem; goto out_unlock_release_sem;
}
/* /*
* We dont have the lock. Look up the PI state (or create it if * We dont have the lock. Look up the PI state (or create it if
...@@ -2289,8 +1898,6 @@ int handle_futex_death(u32 __user *uaddr, struct task_struct *curr, int pi) ...@@ -2289,8 +1898,6 @@ int handle_futex_death(u32 __user *uaddr, struct task_struct *curr, int pi)
* userspace. * userspace.
*/ */
mval = (uval & FUTEX_WAITERS) | FUTEX_OWNER_DIED; mval = (uval & FUTEX_WAITERS) | FUTEX_OWNER_DIED;
/* Also keep the FUTEX_WAITER_REQUEUED flag if set */
mval |= (uval & FUTEX_WAITER_REQUEUED);
nval = futex_atomic_cmpxchg_inatomic(uaddr, uval, mval); nval = futex_atomic_cmpxchg_inatomic(uaddr, uval, mval);
if (nval == -EFAULT) if (nval == -EFAULT)
...@@ -2427,9 +2034,6 @@ long do_futex(u32 __user *uaddr, int op, u32 val, ktime_t *timeout, ...@@ -2427,9 +2034,6 @@ long do_futex(u32 __user *uaddr, int op, u32 val, ktime_t *timeout,
case FUTEX_TRYLOCK_PI: case FUTEX_TRYLOCK_PI:
ret = futex_lock_pi(uaddr, fshared, 0, timeout, 1); ret = futex_lock_pi(uaddr, fshared, 0, timeout, 1);
break; break;
case FUTEX_CMP_REQUEUE_PI:
ret = futex_requeue_pi(uaddr, fshared, uaddr2, val, val2, &val3);
break;
default: default:
ret = -ENOSYS; ret = -ENOSYS;
} }
...@@ -2460,8 +2064,7 @@ asmlinkage long sys_futex(u32 __user *uaddr, int op, u32 val, ...@@ -2460,8 +2064,7 @@ asmlinkage long sys_futex(u32 __user *uaddr, int op, u32 val,
/* /*
* requeue parameter in 'utime' if cmd == FUTEX_REQUEUE. * requeue parameter in 'utime' if cmd == FUTEX_REQUEUE.
*/ */
if (cmd == FUTEX_REQUEUE || cmd == FUTEX_CMP_REQUEUE if (cmd == FUTEX_REQUEUE || cmd == FUTEX_CMP_REQUEUE)
|| cmd == FUTEX_CMP_REQUEUE_PI)
val2 = (u32) (unsigned long) utime; val2 = (u32) (unsigned long) utime;
return do_futex(uaddr, op, val, tp, uaddr2, val2, val3); return do_futex(uaddr, op, val, tp, uaddr2, val2, val3);
......
...@@ -157,8 +157,7 @@ asmlinkage long compat_sys_futex(u32 __user *uaddr, int op, u32 val, ...@@ -157,8 +157,7 @@ asmlinkage long compat_sys_futex(u32 __user *uaddr, int op, u32 val,
t = ktime_add(ktime_get(), t); t = ktime_add(ktime_get(), t);
tp = &t; tp = &t;
} }
if (cmd == FUTEX_REQUEUE || cmd == FUTEX_CMP_REQUEUE if (cmd == FUTEX_REQUEUE || cmd == FUTEX_CMP_REQUEUE)
|| cmd == FUTEX_CMP_REQUEUE_PI)
val2 = (int) (unsigned long) utime; val2 = (int) (unsigned long) utime;
return do_futex(uaddr, op, val, tp, uaddr2, val2, val3); return do_futex(uaddr, op, val, tp, uaddr2, val2, val3);
......
...@@ -56,7 +56,7 @@ ...@@ -56,7 +56,7 @@
* state. * state.
*/ */
void static void
rt_mutex_set_owner(struct rt_mutex *lock, struct task_struct *owner, rt_mutex_set_owner(struct rt_mutex *lock, struct task_struct *owner,
unsigned long mask) unsigned long mask)
{ {
...@@ -80,6 +80,29 @@ static void fixup_rt_mutex_waiters(struct rt_mutex *lock) ...@@ -80,6 +80,29 @@ static void fixup_rt_mutex_waiters(struct rt_mutex *lock)
clear_rt_mutex_waiters(lock); clear_rt_mutex_waiters(lock);
} }
/*
* We can speed up the acquire/release, if the architecture
* supports cmpxchg and if there's no debugging state to be set up
*/
#if defined(__HAVE_ARCH_CMPXCHG) && !defined(CONFIG_DEBUG_RT_MUTEXES)
# define rt_mutex_cmpxchg(l,c,n) (cmpxchg(&l->owner, c, n) == c)
static inline void mark_rt_mutex_waiters(struct rt_mutex *lock)
{
unsigned long owner, *p = (unsigned long *) &lock->owner;
do {
owner = *p;
} while (cmpxchg(p, owner, owner | RT_MUTEX_HAS_WAITERS) != owner);
}
#else
# define rt_mutex_cmpxchg(l,c,n) (0)
static inline void mark_rt_mutex_waiters(struct rt_mutex *lock)
{
lock->owner = (struct task_struct *)
((unsigned long)lock->owner | RT_MUTEX_HAS_WAITERS);
}
#endif
/* /*
* Calculate task priority from the waiter list priority * Calculate task priority from the waiter list priority
* *
...@@ -100,7 +123,7 @@ int rt_mutex_getprio(struct task_struct *task) ...@@ -100,7 +123,7 @@ int rt_mutex_getprio(struct task_struct *task)
* *
* This can be both boosting and unboosting. task->pi_lock must be held. * This can be both boosting and unboosting. task->pi_lock must be held.
*/ */
void __rt_mutex_adjust_prio(struct task_struct *task) static void __rt_mutex_adjust_prio(struct task_struct *task)
{ {
int prio = rt_mutex_getprio(task); int prio = rt_mutex_getprio(task);
...@@ -136,11 +159,11 @@ int max_lock_depth = 1024; ...@@ -136,11 +159,11 @@ int max_lock_depth = 1024;
* Decreases task's usage by one - may thus free the task. * Decreases task's usage by one - may thus free the task.
* Returns 0 or -EDEADLK. * Returns 0 or -EDEADLK.
*/ */
int rt_mutex_adjust_prio_chain(struct task_struct *task, static int rt_mutex_adjust_prio_chain(struct task_struct *task,
int deadlock_detect, int deadlock_detect,
struct rt_mutex *orig_lock, struct rt_mutex *orig_lock,
struct rt_mutex_waiter *orig_waiter, struct rt_mutex_waiter *orig_waiter,
struct task_struct *top_task) struct task_struct *top_task)
{ {
struct rt_mutex *lock; struct rt_mutex *lock;
struct rt_mutex_waiter *waiter, *top_waiter = orig_waiter; struct rt_mutex_waiter *waiter, *top_waiter = orig_waiter;
...@@ -514,8 +537,8 @@ static void wakeup_next_waiter(struct rt_mutex *lock) ...@@ -514,8 +537,8 @@ static void wakeup_next_waiter(struct rt_mutex *lock)
* *
* Must be called with lock->wait_lock held * Must be called with lock->wait_lock held
*/ */
void remove_waiter(struct rt_mutex *lock, static void remove_waiter(struct rt_mutex *lock,
struct rt_mutex_waiter *waiter) struct rt_mutex_waiter *waiter)
{ {
int first = (waiter == rt_mutex_top_waiter(lock)); int first = (waiter == rt_mutex_top_waiter(lock));
struct task_struct *owner = rt_mutex_owner(lock); struct task_struct *owner = rt_mutex_owner(lock);
......
...@@ -112,29 +112,6 @@ static inline unsigned long rt_mutex_owner_pending(struct rt_mutex *lock) ...@@ -112,29 +112,6 @@ static inline unsigned long rt_mutex_owner_pending(struct rt_mutex *lock)
return (unsigned long)lock->owner & RT_MUTEX_OWNER_PENDING; return (unsigned long)lock->owner & RT_MUTEX_OWNER_PENDING;
} }
/*
* We can speed up the acquire/release, if the architecture
* supports cmpxchg and if there's no debugging state to be set up
*/
#if defined(__HAVE_ARCH_CMPXCHG) && !defined(CONFIG_DEBUG_RT_MUTEXES)
# define rt_mutex_cmpxchg(l,c,n) (cmpxchg(&l->owner, c, n) == c)
static inline void mark_rt_mutex_waiters(struct rt_mutex *lock)
{
unsigned long owner, *p = (unsigned long *) &lock->owner;
do {
owner = *p;
} while (cmpxchg(p, owner, owner | RT_MUTEX_HAS_WAITERS) != owner);
}
#else
# define rt_mutex_cmpxchg(l,c,n) (0)
static inline void mark_rt_mutex_waiters(struct rt_mutex *lock)
{
lock->owner = (struct task_struct *)
((unsigned long)lock->owner | RT_MUTEX_HAS_WAITERS);
}
#endif
/* /*
* PI-futex support (proxy locking functions, etc.): * PI-futex support (proxy locking functions, etc.):
*/ */
...@@ -143,15 +120,4 @@ extern void rt_mutex_init_proxy_locked(struct rt_mutex *lock, ...@@ -143,15 +120,4 @@ extern void rt_mutex_init_proxy_locked(struct rt_mutex *lock,
struct task_struct *proxy_owner); struct task_struct *proxy_owner);
extern void rt_mutex_proxy_unlock(struct rt_mutex *lock, extern void rt_mutex_proxy_unlock(struct rt_mutex *lock,
struct task_struct *proxy_owner); struct task_struct *proxy_owner);
extern void rt_mutex_set_owner(struct rt_mutex *lock, struct task_struct *owner,
unsigned long mask);
extern void __rt_mutex_adjust_prio(struct task_struct *task);
extern int rt_mutex_adjust_prio_chain(struct task_struct *task,
int deadlock_detect,
struct rt_mutex *orig_lock,
struct rt_mutex_waiter *orig_waiter,
struct task_struct *top_task);
extern void remove_waiter(struct rt_mutex *lock,
struct rt_mutex_waiter *waiter);
#endif #endif
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册