arm64: start using 'asm goto' for put_user()

This generates noticeably better code since we don't need to test the
error register etc, the exception just jumps to the error handling
directly.

Unlike get_user(), there's no need to worry about old compilers. All
supported compilers support the regular non-output 'asm goto', as
pointed out by Nathan Chancellor.

Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
Linus Torvalds 2024-06-01 15:04:53 -07:00
parent 86a6a68feb
commit 7fd298d4b3
2 changed files with 39 additions and 34 deletions

View File

@ -112,6 +112,9 @@
#define _ASM_EXTABLE_KACCESS_ERR(insn, fixup, err) \ #define _ASM_EXTABLE_KACCESS_ERR(insn, fixup, err) \
_ASM_EXTABLE_KACCESS_ERR_ZERO(insn, fixup, err, wzr) _ASM_EXTABLE_KACCESS_ERR_ZERO(insn, fixup, err, wzr)
#define _ASM_EXTABLE_KACCESS(insn, fixup) \
_ASM_EXTABLE_KACCESS_ERR_ZERO(insn, fixup, wzr, wzr)
#define _ASM_EXTABLE_LOAD_UNALIGNED_ZEROPAD(insn, fixup, data, addr) \ #define _ASM_EXTABLE_LOAD_UNALIGNED_ZEROPAD(insn, fixup, data, addr) \
__DEFINE_ASM_GPR_NUMS \ __DEFINE_ASM_GPR_NUMS \
__ASM_EXTABLE_RAW(#insn, #fixup, \ __ASM_EXTABLE_RAW(#insn, #fixup, \

View File

@ -294,29 +294,28 @@ do { \
} while (0); \ } while (0); \
} while (0) } while (0)
#define __put_mem_asm(store, reg, x, addr, err, type) \ #define __put_mem_asm(store, reg, x, addr, label, type) \
asm volatile( \ asm goto( \
"1: " store " " reg "1, [%2]\n" \ "1: " store " " reg "0, [%1]\n" \
"2:\n" \ "2:\n" \
_ASM_EXTABLE_##type##ACCESS_ERR(1b, 2b, %w0) \ _ASM_EXTABLE_##type##ACCESS(1b, %l2) \
: "+r" (err) \ : : "rZ" (x), "r" (addr) : : label)
: "rZ" (x), "r" (addr))
#define __raw_put_mem(str, x, ptr, err, type) \ #define __raw_put_mem(str, x, ptr, label, type) \
do { \ do { \
__typeof__(*(ptr)) __pu_val = (x); \ __typeof__(*(ptr)) __pu_val = (x); \
switch (sizeof(*(ptr))) { \ switch (sizeof(*(ptr))) { \
case 1: \ case 1: \
__put_mem_asm(str "b", "%w", __pu_val, (ptr), (err), type); \ __put_mem_asm(str "b", "%w", __pu_val, (ptr), label, type); \
break; \ break; \
case 2: \ case 2: \
__put_mem_asm(str "h", "%w", __pu_val, (ptr), (err), type); \ __put_mem_asm(str "h", "%w", __pu_val, (ptr), label, type); \
break; \ break; \
case 4: \ case 4: \
__put_mem_asm(str, "%w", __pu_val, (ptr), (err), type); \ __put_mem_asm(str, "%w", __pu_val, (ptr), label, type); \
break; \ break; \
case 8: \ case 8: \
__put_mem_asm(str, "%x", __pu_val, (ptr), (err), type); \ __put_mem_asm(str, "%x", __pu_val, (ptr), label, type); \
break; \ break; \
default: \ default: \
BUILD_BUG(); \ BUILD_BUG(); \
@ -328,25 +327,34 @@ do { \
* uaccess_ttbr0_disable(). As `x` and `ptr` could contain blocking functions, * uaccess_ttbr0_disable(). As `x` and `ptr` could contain blocking functions,
* we must evaluate these outside of the critical section. * we must evaluate these outside of the critical section.
*/ */
#define __raw_put_user(x, ptr, err) \ #define __raw_put_user(x, ptr, label) \
do { \ do { \
__label__ __rpu_failed; \
__typeof__(*(ptr)) __user *__rpu_ptr = (ptr); \ __typeof__(*(ptr)) __user *__rpu_ptr = (ptr); \
__typeof__(*(ptr)) __rpu_val = (x); \ __typeof__(*(ptr)) __rpu_val = (x); \
__chk_user_ptr(__rpu_ptr); \ __chk_user_ptr(__rpu_ptr); \
\ \
uaccess_ttbr0_enable(); \ do { \
__raw_put_mem("sttr", __rpu_val, __rpu_ptr, err, U); \ uaccess_ttbr0_enable(); \
uaccess_ttbr0_disable(); \ __raw_put_mem("sttr", __rpu_val, __rpu_ptr, __rpu_failed, U); \
uaccess_ttbr0_disable(); \
break; \
__rpu_failed: \
uaccess_ttbr0_disable(); \
goto label; \
} while (0); \
} while (0) } while (0)
#define __put_user_error(x, ptr, err) \ #define __put_user_error(x, ptr, err) \
do { \ do { \
__label__ __pu_failed; \
__typeof__(*(ptr)) __user *__p = (ptr); \ __typeof__(*(ptr)) __user *__p = (ptr); \
might_fault(); \ might_fault(); \
if (access_ok(__p, sizeof(*__p))) { \ if (access_ok(__p, sizeof(*__p))) { \
__p = uaccess_mask_ptr(__p); \ __p = uaccess_mask_ptr(__p); \
__raw_put_user((x), __p, (err)); \ __raw_put_user((x), __p, __pu_failed); \
} else { \ } else { \
__pu_failed: \
(err) = -EFAULT; \ (err) = -EFAULT; \
} \ } \
} while (0) } while (0)
@ -369,15 +377,18 @@ do { \
do { \ do { \
__typeof__(dst) __pkn_dst = (dst); \ __typeof__(dst) __pkn_dst = (dst); \
__typeof__(src) __pkn_src = (src); \ __typeof__(src) __pkn_src = (src); \
int __pkn_err = 0; \
\ \
__mte_enable_tco_async(); \ do { \
__raw_put_mem("str", *((type *)(__pkn_src)), \ __label__ __pkn_err; \
(__force type *)(__pkn_dst), __pkn_err, K); \ __mte_enable_tco_async(); \
__mte_disable_tco_async(); \ __raw_put_mem("str", *((type *)(__pkn_src)), \
\ (__force type *)(__pkn_dst), __pkn_err, K); \
if (unlikely(__pkn_err)) \ __mte_disable_tco_async(); \
break; \
__pkn_err: \
__mte_disable_tco_async(); \
goto err_label; \ goto err_label; \
} while (0); \
} while(0) } while(0)
extern unsigned long __must_check __arch_copy_from_user(void *to, const void __user *from, unsigned long n); extern unsigned long __must_check __arch_copy_from_user(void *to, const void __user *from, unsigned long n);
@ -411,17 +422,8 @@ static __must_check __always_inline bool user_access_begin(const void __user *pt
} }
#define user_access_begin(a,b) user_access_begin(a,b) #define user_access_begin(a,b) user_access_begin(a,b)
#define user_access_end() uaccess_ttbr0_disable() #define user_access_end() uaccess_ttbr0_disable()
#define unsafe_put_user(x, ptr, label) \
/* __raw_put_mem("sttr", x, uaccess_mask_ptr(ptr), label, U)
* The arm64 inline asms should learn abut asm goto, and we should
* teach user_access_begin() about address masking.
*/
#define unsafe_put_user(x, ptr, label) do { \
int __upu_err = 0; \
__raw_put_mem("sttr", x, uaccess_mask_ptr(ptr), __upu_err, U); \
if (__upu_err) goto label; \
} while (0)
#define unsafe_get_user(x, ptr, label) \ #define unsafe_get_user(x, ptr, label) \
__raw_get_mem("ldtr", x, uaccess_mask_ptr(ptr), label, U) __raw_get_mem("ldtr", x, uaccess_mask_ptr(ptr), label, U)