2019-05-19 12:08:55 +00:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-only
|
2008-04-17 18:05:36 +00:00
|
|
|
/*
|
2020-06-09 04:34:11 +00:00
|
|
|
* Access kernel or user memory without faulting.
|
2008-04-17 18:05:36 +00:00
|
|
|
*/
|
2011-10-16 06:01:52 +00:00
|
|
|
#include <linux/export.h>
|
2008-04-17 18:05:36 +00:00
|
|
|
#include <linux/mm.h>
|
2010-10-27 16:29:01 +00:00
|
|
|
#include <linux/uaccess.h>
|
2023-04-10 17:43:44 +00:00
|
|
|
#include <asm/tlb.h>
|
2008-04-17 18:05:36 +00:00
|
|
|
|
2020-06-17 07:37:53 +00:00
|
|
|
bool __weak copy_from_kernel_nofault_allowed(const void *unsafe_src,
|
|
|
|
size_t size)
|
2020-06-09 04:34:27 +00:00
|
|
|
{
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-06-17 07:37:53 +00:00
|
|
|
#define copy_from_kernel_nofault_loop(dst, src, len, type, err_label) \
|
2020-06-09 04:34:58 +00:00
|
|
|
while (len >= sizeof(type)) { \
|
|
|
|
__get_kernel_nofault(dst, src, type, err_label); \
|
|
|
|
dst += sizeof(type); \
|
|
|
|
src += sizeof(type); \
|
|
|
|
len -= sizeof(type); \
|
|
|
|
}
|
|
|
|
|
2020-06-17 07:37:53 +00:00
|
|
|
long copy_from_kernel_nofault(void *dst, const void *src, size_t size)
|
2020-06-09 04:34:58 +00:00
|
|
|
{
|
2021-08-11 07:30:18 +00:00
|
|
|
unsigned long align = 0;
|
|
|
|
|
|
|
|
if (!IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS))
|
|
|
|
align = (unsigned long)dst | (unsigned long)src;
|
|
|
|
|
2020-06-17 07:37:53 +00:00
|
|
|
if (!copy_from_kernel_nofault_allowed(src, size))
|
2020-06-09 04:35:04 +00:00
|
|
|
return -ERANGE;
|
2020-06-09 04:34:58 +00:00
|
|
|
|
|
|
|
pagefault_disable();
|
2021-08-11 07:30:18 +00:00
|
|
|
if (!(align & 7))
|
|
|
|
copy_from_kernel_nofault_loop(dst, src, size, u64, Efault);
|
|
|
|
if (!(align & 3))
|
|
|
|
copy_from_kernel_nofault_loop(dst, src, size, u32, Efault);
|
|
|
|
if (!(align & 1))
|
|
|
|
copy_from_kernel_nofault_loop(dst, src, size, u16, Efault);
|
2020-06-17 07:37:53 +00:00
|
|
|
copy_from_kernel_nofault_loop(dst, src, size, u8, Efault);
|
2020-06-09 04:34:58 +00:00
|
|
|
pagefault_enable();
|
|
|
|
return 0;
|
|
|
|
Efault:
|
|
|
|
pagefault_enable();
|
|
|
|
return -EFAULT;
|
|
|
|
}
|
2020-06-17 07:37:53 +00:00
|
|
|
EXPORT_SYMBOL_GPL(copy_from_kernel_nofault);
|
2020-06-09 04:34:58 +00:00
|
|
|
|
2020-06-17 07:37:53 +00:00
|
|
|
#define copy_to_kernel_nofault_loop(dst, src, len, type, err_label) \
|
2020-06-09 04:34:58 +00:00
|
|
|
while (len >= sizeof(type)) { \
|
|
|
|
__put_kernel_nofault(dst, src, type, err_label); \
|
|
|
|
dst += sizeof(type); \
|
|
|
|
src += sizeof(type); \
|
|
|
|
len -= sizeof(type); \
|
|
|
|
}
|
|
|
|
|
2020-06-17 07:37:53 +00:00
|
|
|
long copy_to_kernel_nofault(void *dst, const void *src, size_t size)
|
2020-06-09 04:34:58 +00:00
|
|
|
{
|
2021-08-11 07:30:18 +00:00
|
|
|
unsigned long align = 0;
|
|
|
|
|
|
|
|
if (!IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS))
|
|
|
|
align = (unsigned long)dst | (unsigned long)src;
|
|
|
|
|
2020-06-09 04:34:58 +00:00
|
|
|
pagefault_disable();
|
2021-08-11 07:30:18 +00:00
|
|
|
if (!(align & 7))
|
|
|
|
copy_to_kernel_nofault_loop(dst, src, size, u64, Efault);
|
|
|
|
if (!(align & 3))
|
|
|
|
copy_to_kernel_nofault_loop(dst, src, size, u32, Efault);
|
|
|
|
if (!(align & 1))
|
|
|
|
copy_to_kernel_nofault_loop(dst, src, size, u16, Efault);
|
2020-06-17 07:37:53 +00:00
|
|
|
copy_to_kernel_nofault_loop(dst, src, size, u8, Efault);
|
2020-06-09 04:34:58 +00:00
|
|
|
pagefault_enable();
|
|
|
|
return 0;
|
|
|
|
Efault:
|
|
|
|
pagefault_enable();
|
|
|
|
return -EFAULT;
|
|
|
|
}
|
|
|
|
|
|
|
|
long strncpy_from_kernel_nofault(char *dst, const void *unsafe_addr, long count)
|
|
|
|
{
|
|
|
|
const void *src = unsafe_addr;
|
|
|
|
|
|
|
|
if (unlikely(count <= 0))
|
|
|
|
return 0;
|
2020-06-17 07:37:53 +00:00
|
|
|
if (!copy_from_kernel_nofault_allowed(unsafe_addr, count))
|
2020-06-09 04:35:04 +00:00
|
|
|
return -ERANGE;
|
2020-06-09 04:34:58 +00:00
|
|
|
|
|
|
|
pagefault_disable();
|
|
|
|
do {
|
|
|
|
__get_kernel_nofault(dst, src, u8, Efault);
|
|
|
|
dst++;
|
|
|
|
src++;
|
|
|
|
} while (dst[-1] && src - unsafe_addr < count);
|
|
|
|
pagefault_enable();
|
|
|
|
|
|
|
|
dst[-1] = '\0';
|
|
|
|
return src - unsafe_addr;
|
|
|
|
Efault:
|
|
|
|
pagefault_enable();
|
2022-11-10 08:56:13 +00:00
|
|
|
dst[0] = '\0';
|
2020-06-09 04:34:58 +00:00
|
|
|
return -EFAULT;
|
|
|
|
}
|
2019-05-15 05:38:18 +00:00
|
|
|
|
2020-06-09 04:34:55 +00:00
|
|
|
/**
|
2020-06-17 07:37:54 +00:00
|
|
|
* copy_from_user_nofault(): safely attempt to read from a user-space location
|
2020-06-09 04:34:55 +00:00
|
|
|
* @dst: pointer to the buffer that shall take the data
|
|
|
|
* @src: address to read from. This must be a user address.
|
|
|
|
* @size: size of the data chunk
|
|
|
|
*
|
|
|
|
* Safely read from user address @src to the buffer at @dst. If a kernel fault
|
|
|
|
* happens, handle that and return -EFAULT.
|
|
|
|
*/
|
2020-06-17 07:37:54 +00:00
|
|
|
long copy_from_user_nofault(void *dst, const void __user *src, size_t size)
|
2020-06-09 04:34:55 +00:00
|
|
|
{
|
|
|
|
long ret = -EFAULT;
|
2023-04-10 17:43:44 +00:00
|
|
|
|
|
|
|
if (!__access_ok(src, size))
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
if (!nmi_uaccess_okay())
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
pagefault_disable();
|
|
|
|
ret = __copy_from_user_inatomic(dst, src, size);
|
|
|
|
pagefault_enable();
|
2020-06-09 04:34:55 +00:00
|
|
|
|
|
|
|
if (ret)
|
|
|
|
return -EFAULT;
|
|
|
|
return 0;
|
|
|
|
}
|
2020-06-17 07:37:54 +00:00
|
|
|
EXPORT_SYMBOL_GPL(copy_from_user_nofault);
|
2020-06-09 04:34:55 +00:00
|
|
|
|
|
|
|
/**
|
2020-06-17 07:37:54 +00:00
|
|
|
* copy_to_user_nofault(): safely attempt to write to a user-space location
|
2020-06-09 04:34:55 +00:00
|
|
|
* @dst: address to write to
|
|
|
|
* @src: pointer to the data that shall be written
|
|
|
|
* @size: size of the data chunk
|
|
|
|
*
|
|
|
|
* Safely write to address @dst from the buffer at @src. If a kernel fault
|
|
|
|
* happens, handle that and return -EFAULT.
|
|
|
|
*/
|
2020-06-17 07:37:54 +00:00
|
|
|
long copy_to_user_nofault(void __user *dst, const void *src, size_t size)
|
2020-06-09 04:34:55 +00:00
|
|
|
{
|
|
|
|
long ret = -EFAULT;
|
|
|
|
|
|
|
|
if (access_ok(dst, size)) {
|
|
|
|
pagefault_disable();
|
|
|
|
ret = __copy_to_user_inatomic(dst, src, size);
|
|
|
|
pagefault_enable();
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ret)
|
|
|
|
return -EFAULT;
|
|
|
|
return 0;
|
|
|
|
}
|
2020-06-17 07:37:54 +00:00
|
|
|
EXPORT_SYMBOL_GPL(copy_to_user_nofault);
|
2020-06-09 04:34:55 +00:00
|
|
|
|
2019-05-15 05:38:18 +00:00
|
|
|
/**
|
2020-06-09 04:34:14 +00:00
|
|
|
* strncpy_from_user_nofault: - Copy a NUL terminated string from unsafe user
|
2019-05-15 05:38:18 +00:00
|
|
|
* address.
|
|
|
|
* @dst: Destination address, in kernel space. This buffer must be at
|
|
|
|
* least @count bytes long.
|
|
|
|
* @unsafe_addr: Unsafe user address.
|
|
|
|
* @count: Maximum number of bytes to copy, including the trailing NUL.
|
|
|
|
*
|
|
|
|
* Copies a NUL-terminated string from unsafe user address to kernel buffer.
|
|
|
|
*
|
|
|
|
* On success, returns the length of the string INCLUDING the trailing NUL.
|
|
|
|
*
|
|
|
|
* If access fails, returns -EFAULT (some data may have been copied
|
|
|
|
* and the trailing NUL added).
|
|
|
|
*
|
|
|
|
* If @count is smaller than the length of the string, copies @count-1 bytes,
|
|
|
|
* sets the last byte of @dst buffer to NUL and returns @count.
|
|
|
|
*/
|
2020-06-09 04:34:14 +00:00
|
|
|
long strncpy_from_user_nofault(char *dst, const void __user *unsafe_addr,
|
2019-05-15 05:38:18 +00:00
|
|
|
long count)
|
|
|
|
{
|
|
|
|
long ret;
|
|
|
|
|
|
|
|
if (unlikely(count <= 0))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
pagefault_disable();
|
|
|
|
ret = strncpy_from_user(dst, unsafe_addr, count);
|
|
|
|
pagefault_enable();
|
|
|
|
|
|
|
|
if (ret >= count) {
|
|
|
|
ret = count;
|
|
|
|
dst[ret - 1] = '\0';
|
|
|
|
} else if (ret > 0) {
|
|
|
|
ret++;
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2020-06-09 04:34:20 +00:00
|
|
|
* strnlen_user_nofault: - Get the size of a user string INCLUDING final NUL.
|
2019-05-15 05:38:18 +00:00
|
|
|
* @unsafe_addr: The string to measure.
|
|
|
|
* @count: Maximum count (including NUL)
|
|
|
|
*
|
|
|
|
* Get the size of a NUL-terminated string in user space without pagefault.
|
|
|
|
*
|
|
|
|
* Returns the size of the string INCLUDING the terminating NUL.
|
|
|
|
*
|
|
|
|
* If the string is too long, returns a number larger than @count. User
|
|
|
|
* has to check the return value against "> count".
|
|
|
|
* On exception (or invalid count), returns 0.
|
|
|
|
*
|
|
|
|
* Unlike strnlen_user, this can be used from IRQ handler etc. because
|
|
|
|
* it disables pagefaults.
|
|
|
|
*/
|
2020-06-09 04:34:20 +00:00
|
|
|
long strnlen_user_nofault(const void __user *unsafe_addr, long count)
|
2019-05-15 05:38:18 +00:00
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
pagefault_disable();
|
|
|
|
ret = strnlen_user(unsafe_addr, count);
|
|
|
|
pagefault_enable();
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
mm: uninline copy_overflow()
While building a small config with CONFIG_CC_OPTIMISE_FOR_SIZE, I ended
up with more than 50 times the following function in vmlinux because GCC
doesn't honor the 'inline' keyword:
c00243bc <copy_overflow>:
c00243bc: 94 21 ff f0 stwu r1,-16(r1)
c00243c0: 7c 85 23 78 mr r5,r4
c00243c4: 7c 64 1b 78 mr r4,r3
c00243c8: 3c 60 c0 62 lis r3,-16286
c00243cc: 7c 08 02 a6 mflr r0
c00243d0: 38 63 5e e5 addi r3,r3,24293
c00243d4: 90 01 00 14 stw r0,20(r1)
c00243d8: 4b ff 82 45 bl c001c61c <__warn_printk>
c00243dc: 0f e0 00 00 twui r0,0
c00243e0: 80 01 00 14 lwz r0,20(r1)
c00243e4: 38 21 00 10 addi r1,r1,16
c00243e8: 7c 08 03 a6 mtlr r0
c00243ec: 4e 80 00 20 blr
With -Winline, GCC tells:
/include/linux/thread_info.h:212:20: warning: inlining failed in call to 'copy_overflow': call is unlikely and code size would grow [-Winline]
copy_overflow() is a non conditional warning called by check_copy_size()
on an error path.
check_copy_size() have to remain inlined in order to benefit from
constant folding, but copy_overflow() is not worth inlining.
Uninline the warning when CONFIG_BUG is selected.
When CONFIG_BUG is not selected, WARN() does nothing so skip it.
This reduces the size of vmlinux by almost 4kbytes.
Link: https://lkml.kernel.org/r/e1723b9cfa924bcefcd41f69d0025b38e4c9364e.1644819985.git.christophe.leroy@csgroup.eu
Signed-off-by: Christophe Leroy <christophe.leroy@csgroup.eu>
Cc: David Laight <David.Laight@ACULAB.COM>
Cc: Anshuman Khandual <anshuman.khandual@arm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2022-03-22 21:47:49 +00:00
|
|
|
|
|
|
|
void __copy_overflow(int size, unsigned long count)
|
|
|
|
{
|
|
|
|
WARN(1, "Buffer overflow detected (%d < %lu)!\n", size, count);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(__copy_overflow);
|