[PATCH RESEND v7 4/6] arm64: futex: refactor futex atomic operation
Yeoreum Yun
yeoreum.yun at arm.com
Tue Sep 16 06:27:37 PDT 2025
Hi Mark,
[...]
> > diff --git a/arch/arm64/include/asm/futex.h b/arch/arm64/include/asm/futex.h
> > index 1d6d9f856ac5..0aeda7ced2c0 100644
> > --- a/arch/arm64/include/asm/futex.h
> > +++ b/arch/arm64/include/asm/futex.h
> > @@ -126,6 +126,60 @@ LSUI_FUTEX_ATOMIC_OP(or, ldtset, al)
> > LSUI_FUTEX_ATOMIC_OP(andnot, ldtclr, al)
> > LSUI_FUTEX_ATOMIC_OP(set, swpt, al)
> >
> > +
> > +#define LSUI_CMPXCHG_HELPER(suffix, start_bit) \
> > +static __always_inline int \
> > +__lsui_cmpxchg_helper_##suffix(u64 __user *uaddr, u32 oldval, u32 newval) \
> > +{ \
> > + int ret = 0; \
> > + u64 oval, nval, tmp; \
> > + \
> > + asm volatile("//__lsui_cmpxchg_helper_" #suffix "\n" \
> > + __LSUI_PREAMBLE \
> > +" prfm pstl1strm, %2\n" \
> > +"1: ldtr %x1, %2\n" \
> > +" mov %x3, %x1\n" \
> > +" bfi %x1, %x5, #" #start_bit ", #32\n" \
> > +" bfi %x3, %x6, #" #start_bit ", #32\n" \
> > +" mov %x4, %x1\n" \
> > +"2: caslt %x1, %x3, %2\n" \
> > +" sub %x1, %x1, %x4\n" \
> > +" cbz %x1, 3f\n" \
> > +" mov %w0, %w7\n" \
> > +"3:\n" \
> > +" dmb ish\n" \
> > +"4:\n" \
> > + _ASM_EXTABLE_UACCESS_ERR(1b, 4b, %w0) \
> > + _ASM_EXTABLE_UACCESS_ERR(2b, 4b, %w0) \
> > + : "+r" (ret), "=&r" (oval), "+Q" (*uaddr), "=&r" (nval), "=&r" (tmp) \
> > + : "r" (oldval), "r" (newval), "Ir" (-EAGAIN) \
> > + : "memory"); \
> > + \
> > + return ret; \
> > +}
> > +
> > +LSUI_CMPXCHG_HELPER(lo, 0)
> > +LSUI_CMPXCHG_HELPER(hi, 32)
> > +
> > +static __always_inline int
> > +__lsui_cmpxchg_helper(u32 __user *uaddr, u32 oldval, u32 newval, u32 *oval)
> > +{
> > + int ret;
> > + unsigned long uaddr_al;
> > +
> > + uaddr_al = ALIGN_DOWN((unsigned long)uaddr, sizeof(u64));
> > +
> > + if (uaddr_al != (unsigned long)uaddr)
> > + ret = __lsui_cmpxchg_helper_hi((u64 __user *)uaddr_al, oldval, newval);
> > + else
> > + ret = __lsui_cmpxchg_helper_lo((u64 __user *)uaddr_al, oldval, newval);
> > +
> > + if (!ret)
> > + *oval = oldval;
> > +
> > + return ret;
> > +}
>
> I think Will expects that you do more of this in C, e.g. have a basic
> user cmpxchg on a 64-bit type, e.g.
>
> /*
> * NOTE: *oldp is NOT updated if a fault is taken.
> */
> static __always_inline int
> user_cmpxchg64_release(u64 __usr *addr, u64 *oldp, u64 new)
> {
> int err = 0;
>
> asm volatile(
> __LSUI_PREAMBLE
> "1: caslt %x[old], %x[new], %[addr]\n"
> "2:\n"
> _ASM_EXTABLE_UACCESS_ERR(1b, 4b, %w0)
> : [addr] "+Q" (addr),
> [old] "+r" (*oldp)
> : [new] "r" (new)
> : "memory"
> );
>
> return err;
> }
>
> That should be the *only* assembly you need to implement.
>
> Atop that, have a wrapper that uses get_user() and that helper above to
> implement the 32-bit user cmpxchg, with all the bit manipulation in C:
Thanks for your suggestion. But small question.
I think it's enough to use usafe_get_user() instead of get_user() in here
since when FEAT_LSUI enabled, it doeesn't need to call
uaccess_ttbr0_enable()/disable().
or Am I missing something?
>
> /*
> * NOTE: *oldp is NOT updated if a fault is taken.
> */
> static int user_cmpxchg32_release(u32 __user *addr, u32 *oldp, u32 new)
> {
> u64 __user *addr64 = PTR_ALIGN_DOWN(addr, sizeof(u64));
> u64 old64, new64;
> int ret;
>
> if (get_user(old64, uaddr64))
> return -EFAULT;
>
> /*
> * TODO: merge '*oldp' into 'old64', and 'new' into 'new64',
> * taking into account endianness and alignment
> */
>
> ret = user_cmpxchg64_release(uaddr64, &old64, new64);
> if (ret)
> return ret;
>
> /*
> * TODO: extract '*oldp' from 'old64', taking into account
> * endianness and alignment.
> */
I think extraction doesn't need since futext_cmpxchg op passes the
"oldval" argument which fetched and passed already.
So when it success, it seems enough to return the passed arguments.
But, I'm not sure which is better whether dividing the cmpxchg64 and
cmpxchge32 interface or make one helper -- user_cmpxchg_release
(__lsui_cmpxchg_helper).
[...]
Thanks.
--
Sincerely,
Yeoreum Yun
More information about the linux-arm-kernel
mailing list