Commit 966cb75f authored by Peter Zijlstra's avatar Peter Zijlstra

futex: Rename: {get,cmpxchg}_futex_value_locked()

In order to prepare introducing these symbols into the global
namespace; rename them:

 s/\<\([^_ ]*\)_futex_value_locked/futex_\1_value_locked/g
Signed-off-by: default avatarPeter Zijlstra (Intel) <peterz@infradead.org>
Signed-off-by: default avatarAndré Almeida <andrealmeid@collabora.com>
Signed-off-by: default avatarPeter Zijlstra (Intel) <peterz@infradead.org>
Reviewed-by: default avatarAndré Almeida <andrealmeid@collabora.com>
Link: https://lore.kernel.org/r/20210923171111.300673-9-andrealmeid@collabora.com
parent eee5a7bc
...@@ -732,7 +732,7 @@ static struct futex_q *futex_top_waiter(struct futex_hash_bucket *hb, ...@@ -732,7 +732,7 @@ static struct futex_q *futex_top_waiter(struct futex_hash_bucket *hb,
return NULL; return NULL;
} }
static int cmpxchg_futex_value_locked(u32 *curval, u32 __user *uaddr, static int futex_cmpxchg_value_locked(u32 *curval, u32 __user *uaddr,
u32 uval, u32 newval) u32 uval, u32 newval)
{ {
int ret; int ret;
...@@ -744,7 +744,7 @@ static int cmpxchg_futex_value_locked(u32 *curval, u32 __user *uaddr, ...@@ -744,7 +744,7 @@ static int cmpxchg_futex_value_locked(u32 *curval, u32 __user *uaddr,
return ret; return ret;
} }
static int get_futex_value_locked(u32 *dest, u32 __user *from) static int futex_get_value_locked(u32 *dest, u32 __user *from)
{ {
int ret; int ret;
...@@ -1070,7 +1070,7 @@ static int attach_to_pi_state(u32 __user *uaddr, u32 uval, ...@@ -1070,7 +1070,7 @@ static int attach_to_pi_state(u32 __user *uaddr, u32 uval,
* still is what we expect it to be, otherwise retry the entire * still is what we expect it to be, otherwise retry the entire
* operation. * operation.
*/ */
if (get_futex_value_locked(&uval2, uaddr)) if (futex_get_value_locked(&uval2, uaddr))
goto out_efault; goto out_efault;
if (uval != uval2) if (uval != uval2)
...@@ -1220,7 +1220,7 @@ static int handle_exit_race(u32 __user *uaddr, u32 uval, ...@@ -1220,7 +1220,7 @@ static int handle_exit_race(u32 __user *uaddr, u32 uval,
* The same logic applies to the case where the exiting task is * The same logic applies to the case where the exiting task is
* already gone. * already gone.
*/ */
if (get_futex_value_locked(&uval2, uaddr)) if (futex_get_value_locked(&uval2, uaddr))
return -EFAULT; return -EFAULT;
/* If the user space value has changed, try again. */ /* If the user space value has changed, try again. */
...@@ -1341,7 +1341,7 @@ static int lock_pi_update_atomic(u32 __user *uaddr, u32 uval, u32 newval) ...@@ -1341,7 +1341,7 @@ static int lock_pi_update_atomic(u32 __user *uaddr, u32 uval, u32 newval)
if (unlikely(should_fail_futex(true))) if (unlikely(should_fail_futex(true)))
return -EFAULT; return -EFAULT;
err = cmpxchg_futex_value_locked(&curval, uaddr, uval, newval); err = futex_cmpxchg_value_locked(&curval, uaddr, uval, newval);
if (unlikely(err)) if (unlikely(err))
return err; return err;
...@@ -1388,7 +1388,7 @@ static int futex_lock_pi_atomic(u32 __user *uaddr, struct futex_hash_bucket *hb, ...@@ -1388,7 +1388,7 @@ static int futex_lock_pi_atomic(u32 __user *uaddr, struct futex_hash_bucket *hb,
* Read the user space value first so we can validate a few * Read the user space value first so we can validate a few
* things before proceeding further. * things before proceeding further.
*/ */
if (get_futex_value_locked(&uval, uaddr)) if (futex_get_value_locked(&uval, uaddr))
return -EFAULT; return -EFAULT;
if (unlikely(should_fail_futex(true))) if (unlikely(should_fail_futex(true)))
...@@ -1559,7 +1559,7 @@ static int wake_futex_pi(u32 __user *uaddr, u32 uval, struct futex_pi_state *pi_ ...@@ -1559,7 +1559,7 @@ static int wake_futex_pi(u32 __user *uaddr, u32 uval, struct futex_pi_state *pi_
goto out_unlock; goto out_unlock;
} }
ret = cmpxchg_futex_value_locked(&curval, uaddr, uval, newval); ret = futex_cmpxchg_value_locked(&curval, uaddr, uval, newval);
if (!ret && (curval != uval)) { if (!ret && (curval != uval)) {
/* /*
* If a unconditional UNLOCK_PI operation (user space did not * If a unconditional UNLOCK_PI operation (user space did not
...@@ -2006,7 +2006,7 @@ futex_proxy_trylock_atomic(u32 __user *pifutex, struct futex_hash_bucket *hb1, ...@@ -2006,7 +2006,7 @@ futex_proxy_trylock_atomic(u32 __user *pifutex, struct futex_hash_bucket *hb1,
u32 curval; u32 curval;
int ret; int ret;
if (get_futex_value_locked(&curval, pifutex)) if (futex_get_value_locked(&curval, pifutex))
return -EFAULT; return -EFAULT;
if (unlikely(should_fail_futex(true))) if (unlikely(should_fail_futex(true)))
...@@ -2182,7 +2182,7 @@ int futex_requeue(u32 __user *uaddr1, unsigned int flags, u32 __user *uaddr2, ...@@ -2182,7 +2182,7 @@ int futex_requeue(u32 __user *uaddr1, unsigned int flags, u32 __user *uaddr2,
if (likely(cmpval != NULL)) { if (likely(cmpval != NULL)) {
u32 curval; u32 curval;
ret = get_futex_value_locked(&curval, uaddr1); ret = futex_get_value_locked(&curval, uaddr1);
if (unlikely(ret)) { if (unlikely(ret)) {
double_unlock_hb(hb1, hb2); double_unlock_hb(hb1, hb2);
...@@ -2628,14 +2628,14 @@ static int __fixup_pi_state_owner(u32 __user *uaddr, struct futex_q *q, ...@@ -2628,14 +2628,14 @@ static int __fixup_pi_state_owner(u32 __user *uaddr, struct futex_q *q,
if (!pi_state->owner) if (!pi_state->owner)
newtid |= FUTEX_OWNER_DIED; newtid |= FUTEX_OWNER_DIED;
err = get_futex_value_locked(&uval, uaddr); err = futex_get_value_locked(&uval, uaddr);
if (err) if (err)
goto handle_err; goto handle_err;
for (;;) { for (;;) {
newval = (uval & FUTEX_OWNER_DIED) | newtid; newval = (uval & FUTEX_OWNER_DIED) | newtid;
err = cmpxchg_futex_value_locked(&curval, uaddr, uval, newval); err = futex_cmpxchg_value_locked(&curval, uaddr, uval, newval);
if (err) if (err)
goto handle_err; goto handle_err;
...@@ -2872,7 +2872,7 @@ static int futex_wait_setup(u32 __user *uaddr, u32 val, unsigned int flags, ...@@ -2872,7 +2872,7 @@ static int futex_wait_setup(u32 __user *uaddr, u32 val, unsigned int flags,
retry_private: retry_private:
*hb = futex_q_lock(q); *hb = futex_q_lock(q);
ret = get_futex_value_locked(&uval, uaddr); ret = futex_get_value_locked(&uval, uaddr);
if (ret) { if (ret) {
futex_q_unlock(*hb); futex_q_unlock(*hb);
...@@ -3250,7 +3250,7 @@ int futex_unlock_pi(u32 __user *uaddr, unsigned int flags) ...@@ -3250,7 +3250,7 @@ int futex_unlock_pi(u32 __user *uaddr, unsigned int flags)
* preserve the WAITERS bit not the OWNER_DIED one. We are the * preserve the WAITERS bit not the OWNER_DIED one. We are the
* owner. * owner.
*/ */
if ((ret = cmpxchg_futex_value_locked(&curval, uaddr, uval, 0))) { if ((ret = futex_cmpxchg_value_locked(&curval, uaddr, uval, 0))) {
spin_unlock(&hb->lock); spin_unlock(&hb->lock);
switch (ret) { switch (ret) {
case -EFAULT: case -EFAULT:
...@@ -3588,7 +3588,7 @@ static int handle_futex_death(u32 __user *uaddr, struct task_struct *curr, ...@@ -3588,7 +3588,7 @@ static int handle_futex_death(u32 __user *uaddr, struct task_struct *curr,
* does not guarantee R/W access. If that fails we * does not guarantee R/W access. If that fails we
* give up and leave the futex locked. * give up and leave the futex locked.
*/ */
if ((err = cmpxchg_futex_value_locked(&nval, uaddr, uval, mval))) { if ((err = futex_cmpxchg_value_locked(&nval, uaddr, uval, mval))) {
switch (err) { switch (err) {
case -EFAULT: case -EFAULT:
if (fault_in_user_writeable(uaddr)) if (fault_in_user_writeable(uaddr))
...@@ -3934,7 +3934,7 @@ static void __init futex_detect_cmpxchg(void) ...@@ -3934,7 +3934,7 @@ static void __init futex_detect_cmpxchg(void)
* implementation, the non-functional ones will return * implementation, the non-functional ones will return
* -ENOSYS. * -ENOSYS.
*/ */
if (cmpxchg_futex_value_locked(&curval, NULL, 0, 0) == -EFAULT) if (futex_cmpxchg_value_locked(&curval, NULL, 0, 0) == -EFAULT)
futex_cmpxchg_enabled = 1; futex_cmpxchg_enabled = 1;
#endif #endif
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment