mirror of
https://github.com/torvalds/linux.git
synced 2024-12-28 05:41:55 +00:00
printk changes for 5.13
-----BEGIN PGP SIGNATURE----- iQIzBAABCAAdFiEESH4wyp42V4tXvYsjUqAMR0iAlPIFAmCIBMIACgkQUqAMR0iA lPIt9w//bbHUN/JsNtLCs/849oExdUn/thVajrD5yELttYZXhdzbXncNdkGX9tlU 4JmExmUoqKYdN6JhSnrcYvckHj7XXZM7pVh9IdzqRh10MEXIQ+7IUHjQc8034Zs/ W4/oZmfMtBjszap+cJ9hvdp9qaJkPz/fRLGlrbjc1K4hhxDa1gGmeD35SKswGltm q6RzX3uRl5JbBrYsLoqb28MGYRHhjf2+Pvndoj+5Nn9FtwPSot6jAkyqY5Y6iJlS W2EsFqOt+Kv7/I93FyQlnXC6Nx7vntmow7knmmGPXDf2BqLb0J8Bxl3fwuzpQoao nZzL/p9GQ4ZXF6y8gRV8+RzPIcftBdayOswEDGH0LzlTkbAe/9Sq9Lo7a4Z8jxHW ro0P+PSRK5Ksm7jvpVmSTg+Nt+XqDA5zA1lAorX1UjsyeDDNF9ndQ4C+ZNhCKo54 y+RDgtAArJMIvsHLQ53ReoOct5NnGVNb8G/r3bIAu+Dn6K3nesr6fP1XG8iduseL yFlLB7w214BQMr2B/C+8lQvj54wWE4lea2+LNvObxC5b8puYj0fEniUxTYP6bcB5 QT+LfTToufYz4US7ggJy6hoEfohifGWVvDHbn9tXmyXotSTHH7pHdYypqY+UO+kl 7BkwzNFCm4qCIKsg8nyJxT2hDOlpcCrQx1dBIjveMqJ0c5+ahXU= =ovSn -----END PGP SIGNATURE----- Merge tag 'printk-for-5.13' of git://git.kernel.org/pub/scm/linux/kernel/git/printk/linux Pull printk updates from Petr Mladek: - Stop synchronizing kernel log buffer readers by logbuf_lock. As a result, the access to the buffer is fully lockless now. Note that printk() itself still uses locks because it tries to flush the messages to the console immediately. Also the per-CPU temporary buffers are still there because they prevent infinite recursion and serialize backtraces from NMI. All this is going to change in the future. - kmsg_dump API rework and cleanup as a side effect of the logbuf_lock removal. - Make bstr_printf() aware that %pf and %pF formats could deference the given pointer. - Show also page flags by %pGp format. - Clarify the documentation for plain pointer printing. - Do not show no_hash_pointers warning multiple times. - Update Senozhatsky email address. - Some clean up. * tag 'printk-for-5.13' of git://git.kernel.org/pub/scm/linux/kernel/git/printk/linux: (24 commits) lib/vsprintf.c: remove leftover 'f' and 'F' cases from bstr_printf() printk: clarify the documentation for plain pointer printing kernel/printk.c: Fixed mundane typos printk: rename vprintk_func to vprintk vsprintf: dump full information of page flags in pGp mm, slub: don't combine pr_err with INFO mm, slub: use pGp to print page flags MAINTAINERS: update Senozhatsky email address lib/vsprintf: do not show no_hash_pointers message multiple times printk: console: remove unnecessary safe buffer usage printk: kmsg_dump: remove _nolock() variants printk: remove logbuf_lock printk: introduce a kmsg_dump iterator printk: kmsg_dumper: remove @active field printk: add syslog_lock printk: use atomic64_t for devkmsg_user.seq printk: use seqcount_latch for clear_seq printk: introduce CONSOLE_LOG_MAX printk: consolidate kmsg_dump_get_buffer/syslog_print_all code printk: refactor kmsg_dump_get_buffer() ...
This commit is contained in:
commit
7f3d08b255
@ -79,7 +79,19 @@ Pointers printed without a specifier extension (i.e unadorned %p) are
|
||||
hashed to prevent leaking information about the kernel memory layout. This
|
||||
has the added benefit of providing a unique identifier. On 64-bit machines
|
||||
the first 32 bits are zeroed. The kernel will print ``(ptrval)`` until it
|
||||
gathers enough entropy. If you *really* want the address see %px below.
|
||||
gathers enough entropy.
|
||||
|
||||
When possible, use specialised modifiers such as %pS or %pB (described below)
|
||||
to avoid the need of providing an unhashed address that has to be interpreted
|
||||
post-hoc. If not possible, and the aim of printing the address is to provide
|
||||
more information for debugging, use %p and boot the kernel with the
|
||||
``no_hash_pointers`` parameter during debugging, which will print all %p
|
||||
addresses unmodified. If you *really* always want the unmodified address, see
|
||||
%px below.
|
||||
|
||||
If (and only if) you are printing addresses as a content of a virtual file in
|
||||
e.g. procfs or sysfs (using e.g. seq_printf(), not printk()) read by a
|
||||
userspace process, use the %pK modifier described below instead of %p or %px.
|
||||
|
||||
Error Pointers
|
||||
--------------
|
||||
@ -139,6 +151,11 @@ For printing kernel pointers which should be hidden from unprivileged
|
||||
users. The behaviour of %pK depends on the kptr_restrict sysctl - see
|
||||
Documentation/admin-guide/sysctl/kernel.rst for more details.
|
||||
|
||||
This modifier is *only* intended when producing content of a file read by
|
||||
userspace from e.g. procfs or sysfs, not for dmesg. Please refer to the
|
||||
section about %p above for discussion about how to manage hashing pointers
|
||||
in printk().
|
||||
|
||||
Unmodified Addresses
|
||||
--------------------
|
||||
|
||||
@ -153,6 +170,13 @@ equivalent to %lx (or %lu). %px is preferred because it is more uniquely
|
||||
grep'able. If in the future we need to modify the way the kernel handles
|
||||
printing pointers we will be better equipped to find the call sites.
|
||||
|
||||
Before using %px, consider if using %p is sufficient together with enabling the
|
||||
``no_hash_pointers`` kernel parameter during debugging sessions (see the %p
|
||||
description above). One valid scenario for %px might be printing information
|
||||
immediately before a panic, which prevents any sensitive information to be
|
||||
exploited anyway, and with %px there would be no need to reproduce the panic
|
||||
with no_hash_pointers.
|
||||
|
||||
Pointer Differences
|
||||
-------------------
|
||||
|
||||
@ -540,7 +564,7 @@ Flags bitfields such as page flags, gfp_flags
|
||||
|
||||
::
|
||||
|
||||
%pGp referenced|uptodate|lru|active|private
|
||||
%pGp referenced|uptodate|lru|active|private|node=0|zone=2|lastcpupid=0x1fffff
|
||||
%pGg GFP_USER|GFP_DMA32|GFP_NOWARN
|
||||
%pGv read|exec|mayread|maywrite|mayexec|denywrite
|
||||
|
||||
|
@ -14535,7 +14535,7 @@ F: kernel/sched/psi.c
|
||||
|
||||
PRINTK
|
||||
M: Petr Mladek <pmladek@suse.com>
|
||||
M: Sergey Senozhatsky <sergey.senozhatsky@gmail.com>
|
||||
M: Sergey Senozhatsky <senozhatsky@chromium.org>
|
||||
R: Steven Rostedt <rostedt@goodmis.org>
|
||||
R: John Ogness <john.ogness@linutronix.de>
|
||||
S: Maintained
|
||||
@ -19388,7 +19388,7 @@ F: drivers/net/vrf.c
|
||||
VSPRINTF
|
||||
M: Petr Mladek <pmladek@suse.com>
|
||||
M: Steven Rostedt <rostedt@goodmis.org>
|
||||
M: Sergey Senozhatsky <sergey.senozhatsky@gmail.com>
|
||||
M: Sergey Senozhatsky <senozhatsky@chromium.org>
|
||||
R: Andy Shevchenko <andriy.shevchenko@linux.intel.com>
|
||||
R: Rasmus Villemoes <linux@rasmusvillemoes.dk>
|
||||
S: Maintained
|
||||
@ -20039,7 +20039,7 @@ F: drivers/staging/media/zoran/
|
||||
ZRAM COMPRESSED RAM BLOCK DEVICE DRVIER
|
||||
M: Minchan Kim <minchan@kernel.org>
|
||||
M: Nitin Gupta <ngupta@vflare.org>
|
||||
R: Sergey Senozhatsky <sergey.senozhatsky.work@gmail.com>
|
||||
R: Sergey Senozhatsky <senozhatsky@chromium.org>
|
||||
L: linux-kernel@vger.kernel.org
|
||||
S: Maintained
|
||||
F: Documentation/admin-guide/blockdev/zram.rst
|
||||
@ -20053,7 +20053,7 @@ F: drivers/tty/serial/zs.*
|
||||
ZSMALLOC COMPRESSED SLAB MEMORY ALLOCATOR
|
||||
M: Minchan Kim <minchan@kernel.org>
|
||||
M: Nitin Gupta <ngupta@vflare.org>
|
||||
R: Sergey Senozhatsky <sergey.senozhatsky.work@gmail.com>
|
||||
R: Sergey Senozhatsky <senozhatsky@chromium.org>
|
||||
L: linux-mm@kvack.org
|
||||
S: Maintained
|
||||
F: Documentation/vm/zsmalloc.rst
|
||||
|
@ -647,6 +647,7 @@ static void oops_to_nvram(struct kmsg_dumper *dumper,
|
||||
{
|
||||
struct oops_log_info *oops_hdr = (struct oops_log_info *)oops_buf;
|
||||
static unsigned int oops_count = 0;
|
||||
static struct kmsg_dump_iter iter;
|
||||
static bool panicking = false;
|
||||
static DEFINE_SPINLOCK(lock);
|
||||
unsigned long flags;
|
||||
@ -681,13 +682,14 @@ static void oops_to_nvram(struct kmsg_dumper *dumper,
|
||||
return;
|
||||
|
||||
if (big_oops_buf) {
|
||||
kmsg_dump_get_buffer(dumper, false,
|
||||
kmsg_dump_rewind(&iter);
|
||||
kmsg_dump_get_buffer(&iter, false,
|
||||
big_oops_buf, big_oops_buf_sz, &text_len);
|
||||
rc = zip_oops(text_len);
|
||||
}
|
||||
if (rc != 0) {
|
||||
kmsg_dump_rewind(dumper);
|
||||
kmsg_dump_get_buffer(dumper, false,
|
||||
kmsg_dump_rewind(&iter);
|
||||
kmsg_dump_get_buffer(&iter, false,
|
||||
oops_data, oops_data_sz, &text_len);
|
||||
err_type = ERR_TYPE_KERNEL_PANIC;
|
||||
oops_hdr->version = cpu_to_be16(OOPS_HDR_VERSION);
|
||||
|
@ -3001,7 +3001,7 @@ print_address(unsigned long addr)
|
||||
static void
|
||||
dump_log_buf(void)
|
||||
{
|
||||
struct kmsg_dumper dumper = { .active = 1 };
|
||||
struct kmsg_dump_iter iter;
|
||||
unsigned char buf[128];
|
||||
size_t len;
|
||||
|
||||
@ -3013,9 +3013,9 @@ dump_log_buf(void)
|
||||
catch_memory_errors = 1;
|
||||
sync();
|
||||
|
||||
kmsg_dump_rewind_nolock(&dumper);
|
||||
kmsg_dump_rewind(&iter);
|
||||
xmon_start_pagination();
|
||||
while (kmsg_dump_get_line_nolock(&dumper, false, buf, sizeof(buf), &len)) {
|
||||
while (kmsg_dump_get_line(&iter, false, buf, sizeof(buf), &len)) {
|
||||
buf[len] = '\0';
|
||||
printf("%s", buf);
|
||||
}
|
||||
|
@ -1,5 +1,6 @@
|
||||
// SPDX-License-Identifier: GPL-2.0
|
||||
#include <linux/kmsg_dump.h>
|
||||
#include <linux/spinlock.h>
|
||||
#include <linux/console.h>
|
||||
#include <linux/string.h>
|
||||
#include <shared/init.h>
|
||||
@ -9,8 +10,11 @@
|
||||
static void kmsg_dumper_stdout(struct kmsg_dumper *dumper,
|
||||
enum kmsg_dump_reason reason)
|
||||
{
|
||||
static struct kmsg_dump_iter iter;
|
||||
static DEFINE_SPINLOCK(lock);
|
||||
static char line[1024];
|
||||
struct console *con;
|
||||
unsigned long flags;
|
||||
size_t len = 0;
|
||||
|
||||
/* only dump kmsg when no console is available */
|
||||
@ -29,11 +33,18 @@ static void kmsg_dumper_stdout(struct kmsg_dumper *dumper,
|
||||
if (con)
|
||||
return;
|
||||
|
||||
if (!spin_trylock_irqsave(&lock, flags))
|
||||
return;
|
||||
|
||||
kmsg_dump_rewind(&iter);
|
||||
|
||||
printf("kmsg_dump:\n");
|
||||
while (kmsg_dump_get_line(dumper, true, line, sizeof(line), &len)) {
|
||||
while (kmsg_dump_get_line(&iter, true, line, sizeof(line), &len)) {
|
||||
line[len] = '\0';
|
||||
printf("%s", line);
|
||||
}
|
||||
|
||||
spin_unlock_irqrestore(&lock, flags);
|
||||
}
|
||||
|
||||
static struct kmsg_dumper kmsg_dumper = {
|
||||
|
@ -1399,6 +1399,7 @@ static irqreturn_t vmbus_percpu_isr(int irq, void *dev_id)
|
||||
static void hv_kmsg_dump(struct kmsg_dumper *dumper,
|
||||
enum kmsg_dump_reason reason)
|
||||
{
|
||||
struct kmsg_dump_iter iter;
|
||||
size_t bytes_written;
|
||||
|
||||
/* We are only interested in panics. */
|
||||
@ -1409,7 +1410,8 @@ static void hv_kmsg_dump(struct kmsg_dumper *dumper,
|
||||
* Write dump contents to the page. No need to synchronize; panic should
|
||||
* be single-threaded.
|
||||
*/
|
||||
kmsg_dump_get_buffer(dumper, false, hv_panic_page, HV_HYP_PAGE_SIZE,
|
||||
kmsg_dump_rewind(&iter);
|
||||
kmsg_dump_get_buffer(&iter, false, hv_panic_page, HV_HYP_PAGE_SIZE,
|
||||
&bytes_written);
|
||||
if (!bytes_written)
|
||||
return;
|
||||
|
@ -52,6 +52,7 @@ static struct mtdoops_context {
|
||||
int nextcount;
|
||||
unsigned long *oops_page_used;
|
||||
|
||||
unsigned long oops_buf_busy;
|
||||
void *oops_buf;
|
||||
} oops_cxt;
|
||||
|
||||
@ -180,6 +181,9 @@ static void mtdoops_write(struct mtdoops_context *cxt, int panic)
|
||||
u32 *hdr;
|
||||
int ret;
|
||||
|
||||
if (test_and_set_bit(0, &cxt->oops_buf_busy))
|
||||
return;
|
||||
|
||||
/* Add mtdoops header to the buffer */
|
||||
hdr = cxt->oops_buf;
|
||||
hdr[0] = cxt->nextcount;
|
||||
@ -190,7 +194,7 @@ static void mtdoops_write(struct mtdoops_context *cxt, int panic)
|
||||
record_size, &retlen, cxt->oops_buf);
|
||||
if (ret == -EOPNOTSUPP) {
|
||||
printk(KERN_ERR "mtdoops: Cannot write from panic without panic_write\n");
|
||||
return;
|
||||
goto out;
|
||||
}
|
||||
} else
|
||||
ret = mtd_write(mtd, cxt->nextpage * record_size,
|
||||
@ -203,6 +207,8 @@ static void mtdoops_write(struct mtdoops_context *cxt, int panic)
|
||||
memset(cxt->oops_buf, 0xff, record_size);
|
||||
|
||||
mtdoops_inc_counter(cxt);
|
||||
out:
|
||||
clear_bit(0, &cxt->oops_buf_busy);
|
||||
}
|
||||
|
||||
static void mtdoops_workfunc_write(struct work_struct *work)
|
||||
@ -271,13 +277,19 @@ static void mtdoops_do_dump(struct kmsg_dumper *dumper,
|
||||
{
|
||||
struct mtdoops_context *cxt = container_of(dumper,
|
||||
struct mtdoops_context, dump);
|
||||
struct kmsg_dump_iter iter;
|
||||
|
||||
/* Only dump oopses if dump_oops is set */
|
||||
if (reason == KMSG_DUMP_OOPS && !dump_oops)
|
||||
return;
|
||||
|
||||
kmsg_dump_get_buffer(dumper, true, cxt->oops_buf + MTDOOPS_HEADER_SIZE,
|
||||
kmsg_dump_rewind(&iter);
|
||||
|
||||
if (test_and_set_bit(0, &cxt->oops_buf_busy))
|
||||
return;
|
||||
kmsg_dump_get_buffer(&iter, true, cxt->oops_buf + MTDOOPS_HEADER_SIZE,
|
||||
record_size - MTDOOPS_HEADER_SIZE, NULL);
|
||||
clear_bit(0, &cxt->oops_buf_busy);
|
||||
|
||||
if (reason != KMSG_DUMP_OOPS) {
|
||||
/* Panics must be written immediately */
|
||||
@ -394,6 +406,7 @@ static int __init mtdoops_init(void)
|
||||
return -ENOMEM;
|
||||
}
|
||||
memset(cxt->oops_buf, 0xff, record_size);
|
||||
cxt->oops_buf_busy = 0;
|
||||
|
||||
INIT_WORK(&cxt->work_erase, mtdoops_workfunc_erase);
|
||||
INIT_WORK(&cxt->work_write, mtdoops_workfunc_write);
|
||||
|
@ -385,6 +385,7 @@ void pstore_record_init(struct pstore_record *record,
|
||||
static void pstore_dump(struct kmsg_dumper *dumper,
|
||||
enum kmsg_dump_reason reason)
|
||||
{
|
||||
struct kmsg_dump_iter iter;
|
||||
unsigned long total = 0;
|
||||
const char *why;
|
||||
unsigned int part = 1;
|
||||
@ -405,6 +406,8 @@ static void pstore_dump(struct kmsg_dumper *dumper,
|
||||
}
|
||||
}
|
||||
|
||||
kmsg_dump_rewind(&iter);
|
||||
|
||||
oopscount++;
|
||||
while (total < kmsg_bytes) {
|
||||
char *dst;
|
||||
@ -435,7 +438,7 @@ static void pstore_dump(struct kmsg_dumper *dumper,
|
||||
dst_size -= header_size;
|
||||
|
||||
/* Write dump contents. */
|
||||
if (!kmsg_dump_get_buffer(dumper, true, dst + header_size,
|
||||
if (!kmsg_dump_get_buffer(&iter, true, dst + header_size,
|
||||
dst_size, &dump_size))
|
||||
break;
|
||||
|
||||
|
@ -29,6 +29,16 @@ enum kmsg_dump_reason {
|
||||
KMSG_DUMP_MAX
|
||||
};
|
||||
|
||||
/**
|
||||
* struct kmsg_dump_iter - iterator for retrieving kernel messages
|
||||
* @cur_seq: Points to the oldest message to dump
|
||||
* @next_seq: Points after the newest message to dump
|
||||
*/
|
||||
struct kmsg_dump_iter {
|
||||
u64 cur_seq;
|
||||
u64 next_seq;
|
||||
};
|
||||
|
||||
/**
|
||||
* struct kmsg_dumper - kernel crash message dumper structure
|
||||
* @list: Entry in the dumper list (private)
|
||||
@ -41,31 +51,19 @@ struct kmsg_dumper {
|
||||
struct list_head list;
|
||||
void (*dump)(struct kmsg_dumper *dumper, enum kmsg_dump_reason reason);
|
||||
enum kmsg_dump_reason max_reason;
|
||||
bool active;
|
||||
bool registered;
|
||||
|
||||
/* private state of the kmsg iterator */
|
||||
u32 cur_idx;
|
||||
u32 next_idx;
|
||||
u64 cur_seq;
|
||||
u64 next_seq;
|
||||
};
|
||||
|
||||
#ifdef CONFIG_PRINTK
|
||||
void kmsg_dump(enum kmsg_dump_reason reason);
|
||||
|
||||
bool kmsg_dump_get_line_nolock(struct kmsg_dumper *dumper, bool syslog,
|
||||
char *line, size_t size, size_t *len);
|
||||
|
||||
bool kmsg_dump_get_line(struct kmsg_dumper *dumper, bool syslog,
|
||||
bool kmsg_dump_get_line(struct kmsg_dump_iter *iter, bool syslog,
|
||||
char *line, size_t size, size_t *len);
|
||||
|
||||
bool kmsg_dump_get_buffer(struct kmsg_dumper *dumper, bool syslog,
|
||||
char *buf, size_t size, size_t *len);
|
||||
bool kmsg_dump_get_buffer(struct kmsg_dump_iter *iter, bool syslog,
|
||||
char *buf, size_t size, size_t *len_out);
|
||||
|
||||
void kmsg_dump_rewind_nolock(struct kmsg_dumper *dumper);
|
||||
|
||||
void kmsg_dump_rewind(struct kmsg_dumper *dumper);
|
||||
void kmsg_dump_rewind(struct kmsg_dump_iter *iter);
|
||||
|
||||
int kmsg_dump_register(struct kmsg_dumper *dumper);
|
||||
|
||||
@ -77,30 +75,19 @@ static inline void kmsg_dump(enum kmsg_dump_reason reason)
|
||||
{
|
||||
}
|
||||
|
||||
static inline bool kmsg_dump_get_line_nolock(struct kmsg_dumper *dumper,
|
||||
bool syslog, const char *line,
|
||||
size_t size, size_t *len)
|
||||
{
|
||||
return false;
|
||||
}
|
||||
|
||||
static inline bool kmsg_dump_get_line(struct kmsg_dumper *dumper, bool syslog,
|
||||
static inline bool kmsg_dump_get_line(struct kmsg_dump_iter *iter, bool syslog,
|
||||
const char *line, size_t size, size_t *len)
|
||||
{
|
||||
return false;
|
||||
}
|
||||
|
||||
static inline bool kmsg_dump_get_buffer(struct kmsg_dumper *dumper, bool syslog,
|
||||
static inline bool kmsg_dump_get_buffer(struct kmsg_dump_iter *iter, bool syslog,
|
||||
char *buf, size_t size, size_t *len)
|
||||
{
|
||||
return false;
|
||||
}
|
||||
|
||||
static inline void kmsg_dump_rewind_nolock(struct kmsg_dumper *dumper)
|
||||
{
|
||||
}
|
||||
|
||||
static inline void kmsg_dump_rewind(struct kmsg_dumper *dumper)
|
||||
static inline void kmsg_dump_rewind(struct kmsg_dump_iter *iter)
|
||||
{
|
||||
}
|
||||
|
||||
|
@ -2077,7 +2077,7 @@ static int kdb_dmesg(int argc, const char **argv)
|
||||
int adjust = 0;
|
||||
int n = 0;
|
||||
int skip = 0;
|
||||
struct kmsg_dumper dumper = { .active = 1 };
|
||||
struct kmsg_dump_iter iter;
|
||||
size_t len;
|
||||
char buf[201];
|
||||
|
||||
@ -2102,8 +2102,8 @@ static int kdb_dmesg(int argc, const char **argv)
|
||||
kdb_set(2, setargs);
|
||||
}
|
||||
|
||||
kmsg_dump_rewind_nolock(&dumper);
|
||||
while (kmsg_dump_get_line_nolock(&dumper, 1, NULL, 0, NULL))
|
||||
kmsg_dump_rewind(&iter);
|
||||
while (kmsg_dump_get_line(&iter, 1, NULL, 0, NULL))
|
||||
n++;
|
||||
|
||||
if (lines < 0) {
|
||||
@ -2135,8 +2135,8 @@ static int kdb_dmesg(int argc, const char **argv)
|
||||
if (skip >= n || skip < 0)
|
||||
return 0;
|
||||
|
||||
kmsg_dump_rewind_nolock(&dumper);
|
||||
while (kmsg_dump_get_line_nolock(&dumper, 1, buf, sizeof(buf), &len)) {
|
||||
kmsg_dump_rewind(&iter);
|
||||
while (kmsg_dump_get_line(&iter, 1, buf, sizeof(buf), &len)) {
|
||||
if (skip) {
|
||||
skip--;
|
||||
continue;
|
||||
|
@ -12,8 +12,6 @@
|
||||
|
||||
#define PRINTK_NMI_CONTEXT_OFFSET 0x010000000
|
||||
|
||||
extern raw_spinlock_t logbuf_lock;
|
||||
|
||||
__printf(4, 0)
|
||||
int vprintk_store(int facility, int level,
|
||||
const struct dev_printk_info *dev_info,
|
||||
@ -21,7 +19,6 @@ int vprintk_store(int facility, int level,
|
||||
|
||||
__printf(1, 0) int vprintk_default(const char *fmt, va_list args);
|
||||
__printf(1, 0) int vprintk_deferred(const char *fmt, va_list args);
|
||||
__printf(1, 0) int vprintk_func(const char *fmt, va_list args);
|
||||
void __printk_safe_enter(void);
|
||||
void __printk_safe_exit(void);
|
||||
|
||||
@ -56,10 +53,8 @@ void defer_console_output(void);
|
||||
|
||||
#else
|
||||
|
||||
__printf(1, 0) int vprintk_func(const char *fmt, va_list args) { return 0; }
|
||||
|
||||
/*
|
||||
* In !PRINTK builds we still export logbuf_lock spin_lock, console_sem
|
||||
* In !PRINTK builds we still export console_sem
|
||||
* semaphore and some of console functions (console_unlock()/etc.), so
|
||||
* printk-safe must preserve the existing local IRQ guarantees.
|
||||
*/
|
||||
|
@ -262,7 +262,7 @@ static void __up_console_sem(unsigned long ip)
|
||||
* definitely not the perfect debug tool (we don't know if _WE_
|
||||
* hold it and are racing, but it helps tracking those weird code
|
||||
* paths in the console code where we end up in places I want
|
||||
* locked without the console sempahore held).
|
||||
* locked without the console semaphore held).
|
||||
*/
|
||||
static int console_locked, console_suspended;
|
||||
|
||||
@ -355,62 +355,50 @@ enum log_flags {
|
||||
LOG_CONT = 8, /* text is a fragment of a continuation line */
|
||||
};
|
||||
|
||||
/*
|
||||
* The logbuf_lock protects kmsg buffer, indices, counters. This can be taken
|
||||
* within the scheduler's rq lock. It must be released before calling
|
||||
* console_unlock() or anything else that might wake up a process.
|
||||
*/
|
||||
DEFINE_RAW_SPINLOCK(logbuf_lock);
|
||||
|
||||
/*
|
||||
* Helper macros to lock/unlock logbuf_lock and switch between
|
||||
* printk-safe/unsafe modes.
|
||||
*/
|
||||
#define logbuf_lock_irq() \
|
||||
do { \
|
||||
printk_safe_enter_irq(); \
|
||||
raw_spin_lock(&logbuf_lock); \
|
||||
} while (0)
|
||||
|
||||
#define logbuf_unlock_irq() \
|
||||
do { \
|
||||
raw_spin_unlock(&logbuf_lock); \
|
||||
printk_safe_exit_irq(); \
|
||||
} while (0)
|
||||
|
||||
#define logbuf_lock_irqsave(flags) \
|
||||
do { \
|
||||
printk_safe_enter_irqsave(flags); \
|
||||
raw_spin_lock(&logbuf_lock); \
|
||||
} while (0)
|
||||
|
||||
#define logbuf_unlock_irqrestore(flags) \
|
||||
do { \
|
||||
raw_spin_unlock(&logbuf_lock); \
|
||||
printk_safe_exit_irqrestore(flags); \
|
||||
} while (0)
|
||||
/* syslog_lock protects syslog_* variables and write access to clear_seq. */
|
||||
static DEFINE_RAW_SPINLOCK(syslog_lock);
|
||||
|
||||
#ifdef CONFIG_PRINTK
|
||||
DECLARE_WAIT_QUEUE_HEAD(log_wait);
|
||||
/* All 3 protected by @syslog_lock. */
|
||||
/* the next printk record to read by syslog(READ) or /proc/kmsg */
|
||||
static u64 syslog_seq;
|
||||
static size_t syslog_partial;
|
||||
static bool syslog_time;
|
||||
|
||||
/* All 3 protected by @console_sem. */
|
||||
/* the next printk record to write to the console */
|
||||
static u64 console_seq;
|
||||
static u64 exclusive_console_stop_seq;
|
||||
static unsigned long console_dropped;
|
||||
|
||||
/* the next printk record to read after the last 'clear' command */
|
||||
static u64 clear_seq;
|
||||
struct latched_seq {
|
||||
seqcount_latch_t latch;
|
||||
u64 val[2];
|
||||
};
|
||||
|
||||
/*
|
||||
* The next printk record to read after the last 'clear' command. There are
|
||||
* two copies (updated with seqcount_latch) so that reads can locklessly
|
||||
* access a valid value. Writers are synchronized by @syslog_lock.
|
||||
*/
|
||||
static struct latched_seq clear_seq = {
|
||||
.latch = SEQCNT_LATCH_ZERO(clear_seq.latch),
|
||||
.val[0] = 0,
|
||||
.val[1] = 0,
|
||||
};
|
||||
|
||||
#ifdef CONFIG_PRINTK_CALLER
|
||||
#define PREFIX_MAX 48
|
||||
#else
|
||||
#define PREFIX_MAX 32
|
||||
#endif
|
||||
#define LOG_LINE_MAX (1024 - PREFIX_MAX)
|
||||
|
||||
/* the maximum size of a formatted record (i.e. with prefix added per line) */
|
||||
#define CONSOLE_LOG_MAX 1024
|
||||
|
||||
/* the maximum size allowed to be reserved for a record */
|
||||
#define LOG_LINE_MAX (CONSOLE_LOG_MAX - PREFIX_MAX)
|
||||
|
||||
#define LOG_LEVEL(v) ((v) & 0x07)
|
||||
#define LOG_FACILITY(v) ((v) >> 3 & 0xff)
|
||||
@ -452,6 +440,31 @@ bool printk_percpu_data_ready(void)
|
||||
return __printk_percpu_data_ready;
|
||||
}
|
||||
|
||||
/* Must be called under syslog_lock. */
|
||||
static void latched_seq_write(struct latched_seq *ls, u64 val)
|
||||
{
|
||||
raw_write_seqcount_latch(&ls->latch);
|
||||
ls->val[0] = val;
|
||||
raw_write_seqcount_latch(&ls->latch);
|
||||
ls->val[1] = val;
|
||||
}
|
||||
|
||||
/* Can be called from any context. */
|
||||
static u64 latched_seq_read_nolock(struct latched_seq *ls)
|
||||
{
|
||||
unsigned int seq;
|
||||
unsigned int idx;
|
||||
u64 val;
|
||||
|
||||
do {
|
||||
seq = raw_read_seqcount_latch(&ls->latch);
|
||||
idx = seq & 0x1;
|
||||
val = ls->val[idx];
|
||||
} while (read_seqcount_latch_retry(&ls->latch, seq));
|
||||
|
||||
return val;
|
||||
}
|
||||
|
||||
/* Return log buffer address */
|
||||
char *log_buf_addr_get(void)
|
||||
{
|
||||
@ -619,7 +632,7 @@ out:
|
||||
|
||||
/* /dev/kmsg - userspace message inject/listen interface */
|
||||
struct devkmsg_user {
|
||||
u64 seq;
|
||||
atomic64_t seq;
|
||||
struct ratelimit_state rs;
|
||||
struct mutex lock;
|
||||
char buf[CONSOLE_EXT_LOG_MAX];
|
||||
@ -719,27 +732,27 @@ static ssize_t devkmsg_read(struct file *file, char __user *buf,
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
logbuf_lock_irq();
|
||||
if (!prb_read_valid(prb, user->seq, r)) {
|
||||
printk_safe_enter_irq();
|
||||
if (!prb_read_valid(prb, atomic64_read(&user->seq), r)) {
|
||||
if (file->f_flags & O_NONBLOCK) {
|
||||
ret = -EAGAIN;
|
||||
logbuf_unlock_irq();
|
||||
printk_safe_exit_irq();
|
||||
goto out;
|
||||
}
|
||||
|
||||
logbuf_unlock_irq();
|
||||
printk_safe_exit_irq();
|
||||
ret = wait_event_interruptible(log_wait,
|
||||
prb_read_valid(prb, user->seq, r));
|
||||
prb_read_valid(prb, atomic64_read(&user->seq), r));
|
||||
if (ret)
|
||||
goto out;
|
||||
logbuf_lock_irq();
|
||||
printk_safe_enter_irq();
|
||||
}
|
||||
|
||||
if (r->info->seq != user->seq) {
|
||||
if (r->info->seq != atomic64_read(&user->seq)) {
|
||||
/* our last seen message is gone, return error and reset */
|
||||
user->seq = r->info->seq;
|
||||
atomic64_set(&user->seq, r->info->seq);
|
||||
ret = -EPIPE;
|
||||
logbuf_unlock_irq();
|
||||
printk_safe_exit_irq();
|
||||
goto out;
|
||||
}
|
||||
|
||||
@ -748,8 +761,8 @@ static ssize_t devkmsg_read(struct file *file, char __user *buf,
|
||||
&r->text_buf[0], r->info->text_len,
|
||||
&r->info->dev_info);
|
||||
|
||||
user->seq = r->info->seq + 1;
|
||||
logbuf_unlock_irq();
|
||||
atomic64_set(&user->seq, r->info->seq + 1);
|
||||
printk_safe_exit_irq();
|
||||
|
||||
if (len > count) {
|
||||
ret = -EINVAL;
|
||||
@ -784,11 +797,11 @@ static loff_t devkmsg_llseek(struct file *file, loff_t offset, int whence)
|
||||
if (offset)
|
||||
return -ESPIPE;
|
||||
|
||||
logbuf_lock_irq();
|
||||
printk_safe_enter_irq();
|
||||
switch (whence) {
|
||||
case SEEK_SET:
|
||||
/* the first record */
|
||||
user->seq = prb_first_valid_seq(prb);
|
||||
atomic64_set(&user->seq, prb_first_valid_seq(prb));
|
||||
break;
|
||||
case SEEK_DATA:
|
||||
/*
|
||||
@ -796,16 +809,16 @@ static loff_t devkmsg_llseek(struct file *file, loff_t offset, int whence)
|
||||
* like issued by 'dmesg -c'. Reading /dev/kmsg itself
|
||||
* changes no global state, and does not clear anything.
|
||||
*/
|
||||
user->seq = clear_seq;
|
||||
atomic64_set(&user->seq, latched_seq_read_nolock(&clear_seq));
|
||||
break;
|
||||
case SEEK_END:
|
||||
/* after the last record */
|
||||
user->seq = prb_next_seq(prb);
|
||||
atomic64_set(&user->seq, prb_next_seq(prb));
|
||||
break;
|
||||
default:
|
||||
ret = -EINVAL;
|
||||
}
|
||||
logbuf_unlock_irq();
|
||||
printk_safe_exit_irq();
|
||||
return ret;
|
||||
}
|
||||
|
||||
@ -820,15 +833,15 @@ static __poll_t devkmsg_poll(struct file *file, poll_table *wait)
|
||||
|
||||
poll_wait(file, &log_wait, wait);
|
||||
|
||||
logbuf_lock_irq();
|
||||
if (prb_read_valid_info(prb, user->seq, &info, NULL)) {
|
||||
printk_safe_enter_irq();
|
||||
if (prb_read_valid_info(prb, atomic64_read(&user->seq), &info, NULL)) {
|
||||
/* return error when data has vanished underneath us */
|
||||
if (info.seq != user->seq)
|
||||
if (info.seq != atomic64_read(&user->seq))
|
||||
ret = EPOLLIN|EPOLLRDNORM|EPOLLERR|EPOLLPRI;
|
||||
else
|
||||
ret = EPOLLIN|EPOLLRDNORM;
|
||||
}
|
||||
logbuf_unlock_irq();
|
||||
printk_safe_exit_irq();
|
||||
|
||||
return ret;
|
||||
}
|
||||
@ -861,9 +874,9 @@ static int devkmsg_open(struct inode *inode, struct file *file)
|
||||
prb_rec_init_rd(&user->record, &user->info,
|
||||
&user->text_buf[0], sizeof(user->text_buf));
|
||||
|
||||
logbuf_lock_irq();
|
||||
user->seq = prb_first_valid_seq(prb);
|
||||
logbuf_unlock_irq();
|
||||
printk_safe_enter_irq();
|
||||
atomic64_set(&user->seq, prb_first_valid_seq(prb));
|
||||
printk_safe_exit_irq();
|
||||
|
||||
file->private_data = user;
|
||||
return 0;
|
||||
@ -955,6 +968,9 @@ void log_buf_vmcoreinfo_setup(void)
|
||||
|
||||
VMCOREINFO_SIZE(atomic_long_t);
|
||||
VMCOREINFO_TYPE_OFFSET(atomic_long_t, counter);
|
||||
|
||||
VMCOREINFO_STRUCT_SIZE(latched_seq);
|
||||
VMCOREINFO_OFFSET(latched_seq, val);
|
||||
}
|
||||
#endif
|
||||
|
||||
@ -1421,6 +1437,50 @@ static size_t get_record_print_text_size(struct printk_info *info,
|
||||
return ((prefix_len * line_count) + info->text_len + 1);
|
||||
}
|
||||
|
||||
/*
|
||||
* Beginning with @start_seq, find the first record where it and all following
|
||||
* records up to (but not including) @max_seq fit into @size.
|
||||
*
|
||||
* @max_seq is simply an upper bound and does not need to exist. If the caller
|
||||
* does not require an upper bound, -1 can be used for @max_seq.
|
||||
*/
|
||||
static u64 find_first_fitting_seq(u64 start_seq, u64 max_seq, size_t size,
|
||||
bool syslog, bool time)
|
||||
{
|
||||
struct printk_info info;
|
||||
unsigned int line_count;
|
||||
size_t len = 0;
|
||||
u64 seq;
|
||||
|
||||
/* Determine the size of the records up to @max_seq. */
|
||||
prb_for_each_info(start_seq, prb, seq, &info, &line_count) {
|
||||
if (info.seq >= max_seq)
|
||||
break;
|
||||
len += get_record_print_text_size(&info, line_count, syslog, time);
|
||||
}
|
||||
|
||||
/*
|
||||
* Adjust the upper bound for the next loop to avoid subtracting
|
||||
* lengths that were never added.
|
||||
*/
|
||||
if (seq < max_seq)
|
||||
max_seq = seq;
|
||||
|
||||
/*
|
||||
* Move first record forward until length fits into the buffer. Ignore
|
||||
* newest messages that were not counted in the above cycle. Messages
|
||||
* might appear and get lost in the meantime. This is a best effort
|
||||
* that prevents an infinite loop that could occur with a retry.
|
||||
*/
|
||||
prb_for_each_info(start_seq, prb, seq, &info, &line_count) {
|
||||
if (len <= size || info.seq >= max_seq)
|
||||
break;
|
||||
len -= get_record_print_text_size(&info, line_count, syslog, time);
|
||||
}
|
||||
|
||||
return seq;
|
||||
}
|
||||
|
||||
static int syslog_print(char __user *buf, int size)
|
||||
{
|
||||
struct printk_info info;
|
||||
@ -1428,19 +1488,21 @@ static int syslog_print(char __user *buf, int size)
|
||||
char *text;
|
||||
int len = 0;
|
||||
|
||||
text = kmalloc(LOG_LINE_MAX + PREFIX_MAX, GFP_KERNEL);
|
||||
text = kmalloc(CONSOLE_LOG_MAX, GFP_KERNEL);
|
||||
if (!text)
|
||||
return -ENOMEM;
|
||||
|
||||
prb_rec_init_rd(&r, &info, text, LOG_LINE_MAX + PREFIX_MAX);
|
||||
prb_rec_init_rd(&r, &info, text, CONSOLE_LOG_MAX);
|
||||
|
||||
while (size > 0) {
|
||||
size_t n;
|
||||
size_t skip;
|
||||
|
||||
logbuf_lock_irq();
|
||||
printk_safe_enter_irq();
|
||||
raw_spin_lock(&syslog_lock);
|
||||
if (!prb_read_valid(prb, syslog_seq, &r)) {
|
||||
logbuf_unlock_irq();
|
||||
raw_spin_unlock(&syslog_lock);
|
||||
printk_safe_exit_irq();
|
||||
break;
|
||||
}
|
||||
if (r.info->seq != syslog_seq) {
|
||||
@ -1469,7 +1531,8 @@ static int syslog_print(char __user *buf, int size)
|
||||
syslog_partial += n;
|
||||
} else
|
||||
n = 0;
|
||||
logbuf_unlock_irq();
|
||||
raw_spin_unlock(&syslog_lock);
|
||||
printk_safe_exit_irq();
|
||||
|
||||
if (!n)
|
||||
break;
|
||||
@ -1492,34 +1555,26 @@ static int syslog_print(char __user *buf, int size)
|
||||
static int syslog_print_all(char __user *buf, int size, bool clear)
|
||||
{
|
||||
struct printk_info info;
|
||||
unsigned int line_count;
|
||||
struct printk_record r;
|
||||
char *text;
|
||||
int len = 0;
|
||||
u64 seq;
|
||||
bool time;
|
||||
|
||||
text = kmalloc(LOG_LINE_MAX + PREFIX_MAX, GFP_KERNEL);
|
||||
text = kmalloc(CONSOLE_LOG_MAX, GFP_KERNEL);
|
||||
if (!text)
|
||||
return -ENOMEM;
|
||||
|
||||
time = printk_time;
|
||||
logbuf_lock_irq();
|
||||
printk_safe_enter_irq();
|
||||
/*
|
||||
* Find first record that fits, including all following records,
|
||||
* into the user-provided buffer for this dump.
|
||||
*/
|
||||
prb_for_each_info(clear_seq, prb, seq, &info, &line_count)
|
||||
len += get_record_print_text_size(&info, line_count, true, time);
|
||||
seq = find_first_fitting_seq(latched_seq_read_nolock(&clear_seq), -1,
|
||||
size, true, time);
|
||||
|
||||
/* move first record forward until length fits into the buffer */
|
||||
prb_for_each_info(clear_seq, prb, seq, &info, &line_count) {
|
||||
if (len <= size)
|
||||
break;
|
||||
len -= get_record_print_text_size(&info, line_count, true, time);
|
||||
}
|
||||
|
||||
prb_rec_init_rd(&r, &info, text, LOG_LINE_MAX + PREFIX_MAX);
|
||||
prb_rec_init_rd(&r, &info, text, CONSOLE_LOG_MAX);
|
||||
|
||||
len = 0;
|
||||
prb_for_each_record(seq, prb, seq, &r) {
|
||||
@ -1532,20 +1587,23 @@ static int syslog_print_all(char __user *buf, int size, bool clear)
|
||||
break;
|
||||
}
|
||||
|
||||
logbuf_unlock_irq();
|
||||
printk_safe_exit_irq();
|
||||
if (copy_to_user(buf + len, text, textlen))
|
||||
len = -EFAULT;
|
||||
else
|
||||
len += textlen;
|
||||
logbuf_lock_irq();
|
||||
printk_safe_enter_irq();
|
||||
|
||||
if (len < 0)
|
||||
break;
|
||||
}
|
||||
|
||||
if (clear)
|
||||
clear_seq = seq;
|
||||
logbuf_unlock_irq();
|
||||
if (clear) {
|
||||
raw_spin_lock(&syslog_lock);
|
||||
latched_seq_write(&clear_seq, seq);
|
||||
raw_spin_unlock(&syslog_lock);
|
||||
}
|
||||
printk_safe_exit_irq();
|
||||
|
||||
kfree(text);
|
||||
return len;
|
||||
@ -1553,9 +1611,23 @@ static int syslog_print_all(char __user *buf, int size, bool clear)
|
||||
|
||||
static void syslog_clear(void)
|
||||
{
|
||||
logbuf_lock_irq();
|
||||
clear_seq = prb_next_seq(prb);
|
||||
logbuf_unlock_irq();
|
||||
printk_safe_enter_irq();
|
||||
raw_spin_lock(&syslog_lock);
|
||||
latched_seq_write(&clear_seq, prb_next_seq(prb));
|
||||
raw_spin_unlock(&syslog_lock);
|
||||
printk_safe_exit_irq();
|
||||
}
|
||||
|
||||
/* Return a consistent copy of @syslog_seq. */
|
||||
static u64 read_syslog_seq_irq(void)
|
||||
{
|
||||
u64 seq;
|
||||
|
||||
raw_spin_lock_irq(&syslog_lock);
|
||||
seq = syslog_seq;
|
||||
raw_spin_unlock_irq(&syslog_lock);
|
||||
|
||||
return seq;
|
||||
}
|
||||
|
||||
int do_syslog(int type, char __user *buf, int len, int source)
|
||||
@ -1581,8 +1653,9 @@ int do_syslog(int type, char __user *buf, int len, int source)
|
||||
return 0;
|
||||
if (!access_ok(buf, len))
|
||||
return -EFAULT;
|
||||
|
||||
error = wait_event_interruptible(log_wait,
|
||||
prb_read_valid(prb, syslog_seq, NULL));
|
||||
prb_read_valid(prb, read_syslog_seq_irq(), NULL));
|
||||
if (error)
|
||||
return error;
|
||||
error = syslog_print(buf, len);
|
||||
@ -1630,10 +1703,12 @@ int do_syslog(int type, char __user *buf, int len, int source)
|
||||
break;
|
||||
/* Number of chars in the log buffer */
|
||||
case SYSLOG_ACTION_SIZE_UNREAD:
|
||||
logbuf_lock_irq();
|
||||
printk_safe_enter_irq();
|
||||
raw_spin_lock(&syslog_lock);
|
||||
if (!prb_read_valid_info(prb, syslog_seq, &info, NULL)) {
|
||||
/* No unread messages. */
|
||||
logbuf_unlock_irq();
|
||||
raw_spin_unlock(&syslog_lock);
|
||||
printk_safe_exit_irq();
|
||||
return 0;
|
||||
}
|
||||
if (info.seq != syslog_seq) {
|
||||
@ -1661,7 +1736,8 @@ int do_syslog(int type, char __user *buf, int len, int source)
|
||||
}
|
||||
error -= syslog_partial;
|
||||
}
|
||||
logbuf_unlock_irq();
|
||||
raw_spin_unlock(&syslog_lock);
|
||||
printk_safe_exit_irq();
|
||||
break;
|
||||
/* Size of the log buffer */
|
||||
case SYSLOG_ACTION_SIZE_BUFFER:
|
||||
@ -2104,12 +2180,6 @@ asmlinkage int vprintk_emit(int facility, int level,
|
||||
}
|
||||
EXPORT_SYMBOL(vprintk_emit);
|
||||
|
||||
asmlinkage int vprintk(const char *fmt, va_list args)
|
||||
{
|
||||
return vprintk_func(fmt, args);
|
||||
}
|
||||
EXPORT_SYMBOL(vprintk);
|
||||
|
||||
int vprintk_default(const char *fmt, va_list args)
|
||||
{
|
||||
return vprintk_emit(0, LOGLEVEL_DEFAULT, NULL, fmt, args);
|
||||
@ -2143,7 +2213,7 @@ asmlinkage __visible int printk(const char *fmt, ...)
|
||||
int r;
|
||||
|
||||
va_start(args, fmt);
|
||||
r = vprintk_func(fmt, args);
|
||||
r = vprintk(fmt, args);
|
||||
va_end(args);
|
||||
|
||||
return r;
|
||||
@ -2152,8 +2222,7 @@ EXPORT_SYMBOL(printk);
|
||||
|
||||
#else /* CONFIG_PRINTK */
|
||||
|
||||
#define LOG_LINE_MAX 0
|
||||
#define PREFIX_MAX 0
|
||||
#define CONSOLE_LOG_MAX 0
|
||||
#define printk_time false
|
||||
|
||||
#define prb_read_valid(rb, seq, r) false
|
||||
@ -2262,7 +2331,7 @@ static int __init console_setup(char *str)
|
||||
/*
|
||||
* console="" or console=null have been suggested as a way to
|
||||
* disable console output. Use ttynull that has been created
|
||||
* for exacly this purpose.
|
||||
* for exactly this purpose.
|
||||
*/
|
||||
if (str[0] == 0 || strcmp(str, "null") == 0) {
|
||||
__add_preferred_console("ttynull", 0, NULL, NULL, true);
|
||||
@ -2471,7 +2540,7 @@ static inline int can_use_console(void)
|
||||
void console_unlock(void)
|
||||
{
|
||||
static char ext_text[CONSOLE_EXT_LOG_MAX];
|
||||
static char text[LOG_LINE_MAX + PREFIX_MAX];
|
||||
static char text[CONSOLE_LOG_MAX];
|
||||
unsigned long flags;
|
||||
bool do_cond_resched, retry;
|
||||
struct printk_info info;
|
||||
@ -2518,7 +2587,6 @@ again:
|
||||
size_t len;
|
||||
|
||||
printk_safe_enter_irqsave(flags);
|
||||
raw_spin_lock(&logbuf_lock);
|
||||
skip:
|
||||
if (!prb_read_valid(prb, console_seq, &r))
|
||||
break;
|
||||
@ -2562,7 +2630,6 @@ skip:
|
||||
console_msg_format & MSG_FORMAT_SYSLOG,
|
||||
printk_time);
|
||||
console_seq++;
|
||||
raw_spin_unlock(&logbuf_lock);
|
||||
|
||||
/*
|
||||
* While actively printing out messages, if another printk()
|
||||
@ -2589,8 +2656,6 @@ skip:
|
||||
|
||||
console_locked = 0;
|
||||
|
||||
raw_spin_unlock(&logbuf_lock);
|
||||
|
||||
up_console_sem();
|
||||
|
||||
/*
|
||||
@ -2599,9 +2664,7 @@ skip:
|
||||
* there's a new owner and the console_unlock() from them will do the
|
||||
* flush, no worries.
|
||||
*/
|
||||
raw_spin_lock(&logbuf_lock);
|
||||
retry = prb_read_valid(prb, console_seq, NULL);
|
||||
raw_spin_unlock(&logbuf_lock);
|
||||
printk_safe_exit_irqrestore(flags);
|
||||
|
||||
if (retry && console_trylock())
|
||||
@ -2668,9 +2731,9 @@ void console_flush_on_panic(enum con_flush_mode mode)
|
||||
if (mode == CONSOLE_REPLAY_ALL) {
|
||||
unsigned long flags;
|
||||
|
||||
logbuf_lock_irqsave(flags);
|
||||
printk_safe_enter_irqsave(flags);
|
||||
console_seq = prb_first_valid_seq(prb);
|
||||
logbuf_unlock_irqrestore(flags);
|
||||
printk_safe_exit_irqrestore(flags);
|
||||
}
|
||||
console_unlock();
|
||||
}
|
||||
@ -2898,9 +2961,7 @@ void register_console(struct console *newcon)
|
||||
/*
|
||||
* console_unlock(); will print out the buffered messages
|
||||
* for us.
|
||||
*/
|
||||
logbuf_lock_irqsave(flags);
|
||||
/*
|
||||
*
|
||||
* We're about to replay the log buffer. Only do this to the
|
||||
* just-registered console to avoid excessive message spam to
|
||||
* the already-registered consoles.
|
||||
@ -2911,8 +2972,11 @@ void register_console(struct console *newcon)
|
||||
*/
|
||||
exclusive_console = newcon;
|
||||
exclusive_console_stop_seq = console_seq;
|
||||
|
||||
/* Get a consistent copy of @syslog_seq. */
|
||||
raw_spin_lock_irqsave(&syslog_lock, flags);
|
||||
console_seq = syslog_seq;
|
||||
logbuf_unlock_irqrestore(flags);
|
||||
raw_spin_unlock_irqrestore(&syslog_lock, flags);
|
||||
}
|
||||
console_unlock();
|
||||
console_sysfs_notify();
|
||||
@ -3042,7 +3106,7 @@ void __init console_init(void)
|
||||
*
|
||||
* To mitigate this problem somewhat, only unregister consoles whose memory
|
||||
* intersects with the init section. Note that all other boot consoles will
|
||||
* get unregistred when the real preferred console is registered.
|
||||
* get unregistered when the real preferred console is registered.
|
||||
*/
|
||||
static int __init printk_late_init(void)
|
||||
{
|
||||
@ -3276,7 +3340,6 @@ EXPORT_SYMBOL_GPL(kmsg_dump_reason_str);
|
||||
void kmsg_dump(enum kmsg_dump_reason reason)
|
||||
{
|
||||
struct kmsg_dumper *dumper;
|
||||
unsigned long flags;
|
||||
|
||||
rcu_read_lock();
|
||||
list_for_each_entry_rcu(dumper, &dump_list, list) {
|
||||
@ -3293,26 +3356,15 @@ void kmsg_dump(enum kmsg_dump_reason reason)
|
||||
if (reason > max_reason)
|
||||
continue;
|
||||
|
||||
/* initialize iterator with data about the stored records */
|
||||
dumper->active = true;
|
||||
|
||||
logbuf_lock_irqsave(flags);
|
||||
dumper->cur_seq = clear_seq;
|
||||
dumper->next_seq = prb_next_seq(prb);
|
||||
logbuf_unlock_irqrestore(flags);
|
||||
|
||||
/* invoke dumper which will iterate over records */
|
||||
dumper->dump(dumper, reason);
|
||||
|
||||
/* reset iterator */
|
||||
dumper->active = false;
|
||||
}
|
||||
rcu_read_unlock();
|
||||
}
|
||||
|
||||
/**
|
||||
* kmsg_dump_get_line_nolock - retrieve one kmsg log line (unlocked version)
|
||||
* @dumper: registered kmsg dumper
|
||||
* kmsg_dump_get_line - retrieve one kmsg log line
|
||||
* @iter: kmsg dump iterator
|
||||
* @syslog: include the "<4>" prefixes
|
||||
* @line: buffer to copy the line to
|
||||
* @size: maximum size of the buffer
|
||||
@ -3326,30 +3378,31 @@ void kmsg_dump(enum kmsg_dump_reason reason)
|
||||
*
|
||||
* A return value of FALSE indicates that there are no more records to
|
||||
* read.
|
||||
*
|
||||
* The function is similar to kmsg_dump_get_line(), but grabs no locks.
|
||||
*/
|
||||
bool kmsg_dump_get_line_nolock(struct kmsg_dumper *dumper, bool syslog,
|
||||
char *line, size_t size, size_t *len)
|
||||
bool kmsg_dump_get_line(struct kmsg_dump_iter *iter, bool syslog,
|
||||
char *line, size_t size, size_t *len)
|
||||
{
|
||||
u64 min_seq = latched_seq_read_nolock(&clear_seq);
|
||||
struct printk_info info;
|
||||
unsigned int line_count;
|
||||
struct printk_record r;
|
||||
unsigned long flags;
|
||||
size_t l = 0;
|
||||
bool ret = false;
|
||||
|
||||
prb_rec_init_rd(&r, &info, line, size);
|
||||
if (iter->cur_seq < min_seq)
|
||||
iter->cur_seq = min_seq;
|
||||
|
||||
if (!dumper->active)
|
||||
goto out;
|
||||
printk_safe_enter_irqsave(flags);
|
||||
prb_rec_init_rd(&r, &info, line, size);
|
||||
|
||||
/* Read text or count text lines? */
|
||||
if (line) {
|
||||
if (!prb_read_valid(prb, dumper->cur_seq, &r))
|
||||
if (!prb_read_valid(prb, iter->cur_seq, &r))
|
||||
goto out;
|
||||
l = record_print_text(&r, syslog, printk_time);
|
||||
} else {
|
||||
if (!prb_read_valid_info(prb, dumper->cur_seq,
|
||||
if (!prb_read_valid_info(prb, iter->cur_seq,
|
||||
&info, &line_count)) {
|
||||
goto out;
|
||||
}
|
||||
@ -3358,52 +3411,23 @@ bool kmsg_dump_get_line_nolock(struct kmsg_dumper *dumper, bool syslog,
|
||||
|
||||
}
|
||||
|
||||
dumper->cur_seq = r.info->seq + 1;
|
||||
iter->cur_seq = r.info->seq + 1;
|
||||
ret = true;
|
||||
out:
|
||||
printk_safe_exit_irqrestore(flags);
|
||||
if (len)
|
||||
*len = l;
|
||||
return ret;
|
||||
}
|
||||
|
||||
/**
|
||||
* kmsg_dump_get_line - retrieve one kmsg log line
|
||||
* @dumper: registered kmsg dumper
|
||||
* @syslog: include the "<4>" prefixes
|
||||
* @line: buffer to copy the line to
|
||||
* @size: maximum size of the buffer
|
||||
* @len: length of line placed into buffer
|
||||
*
|
||||
* Start at the beginning of the kmsg buffer, with the oldest kmsg
|
||||
* record, and copy one record into the provided buffer.
|
||||
*
|
||||
* Consecutive calls will return the next available record moving
|
||||
* towards the end of the buffer with the youngest messages.
|
||||
*
|
||||
* A return value of FALSE indicates that there are no more records to
|
||||
* read.
|
||||
*/
|
||||
bool kmsg_dump_get_line(struct kmsg_dumper *dumper, bool syslog,
|
||||
char *line, size_t size, size_t *len)
|
||||
{
|
||||
unsigned long flags;
|
||||
bool ret;
|
||||
|
||||
logbuf_lock_irqsave(flags);
|
||||
ret = kmsg_dump_get_line_nolock(dumper, syslog, line, size, len);
|
||||
logbuf_unlock_irqrestore(flags);
|
||||
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(kmsg_dump_get_line);
|
||||
|
||||
/**
|
||||
* kmsg_dump_get_buffer - copy kmsg log lines
|
||||
* @dumper: registered kmsg dumper
|
||||
* @iter: kmsg dump iterator
|
||||
* @syslog: include the "<4>" prefixes
|
||||
* @buf: buffer to copy the line to
|
||||
* @size: maximum size of the buffer
|
||||
* @len: length of line placed into buffer
|
||||
* @len_out: length of line placed into buffer
|
||||
*
|
||||
* Start at the end of the kmsg buffer and fill the provided buffer
|
||||
* with as many of the *youngest* kmsg records that fit into it.
|
||||
@ -3416,115 +3440,93 @@ EXPORT_SYMBOL_GPL(kmsg_dump_get_line);
|
||||
* A return value of FALSE indicates that there are no more records to
|
||||
* read.
|
||||
*/
|
||||
bool kmsg_dump_get_buffer(struct kmsg_dumper *dumper, bool syslog,
|
||||
char *buf, size_t size, size_t *len)
|
||||
bool kmsg_dump_get_buffer(struct kmsg_dump_iter *iter, bool syslog,
|
||||
char *buf, size_t size, size_t *len_out)
|
||||
{
|
||||
u64 min_seq = latched_seq_read_nolock(&clear_seq);
|
||||
struct printk_info info;
|
||||
unsigned int line_count;
|
||||
struct printk_record r;
|
||||
unsigned long flags;
|
||||
u64 seq;
|
||||
u64 next_seq;
|
||||
size_t l = 0;
|
||||
size_t len = 0;
|
||||
bool ret = false;
|
||||
bool time = printk_time;
|
||||
|
||||
prb_rec_init_rd(&r, &info, buf, size);
|
||||
|
||||
if (!dumper->active || !buf || !size)
|
||||
if (!buf || !size)
|
||||
goto out;
|
||||
|
||||
logbuf_lock_irqsave(flags);
|
||||
if (prb_read_valid_info(prb, dumper->cur_seq, &info, NULL)) {
|
||||
if (info.seq != dumper->cur_seq) {
|
||||
if (iter->cur_seq < min_seq)
|
||||
iter->cur_seq = min_seq;
|
||||
|
||||
printk_safe_enter_irqsave(flags);
|
||||
if (prb_read_valid_info(prb, iter->cur_seq, &info, NULL)) {
|
||||
if (info.seq != iter->cur_seq) {
|
||||
/* messages are gone, move to first available one */
|
||||
dumper->cur_seq = info.seq;
|
||||
iter->cur_seq = info.seq;
|
||||
}
|
||||
}
|
||||
|
||||
/* last entry */
|
||||
if (dumper->cur_seq >= dumper->next_seq) {
|
||||
logbuf_unlock_irqrestore(flags);
|
||||
if (iter->cur_seq >= iter->next_seq) {
|
||||
printk_safe_exit_irqrestore(flags);
|
||||
goto out;
|
||||
}
|
||||
|
||||
/* calculate length of entire buffer */
|
||||
seq = dumper->cur_seq;
|
||||
while (prb_read_valid_info(prb, seq, &info, &line_count)) {
|
||||
if (r.info->seq >= dumper->next_seq)
|
||||
break;
|
||||
l += get_record_print_text_size(&info, line_count, syslog, time);
|
||||
seq = r.info->seq + 1;
|
||||
}
|
||||
/*
|
||||
* Find first record that fits, including all following records,
|
||||
* into the user-provided buffer for this dump. Pass in size-1
|
||||
* because this function (by way of record_print_text()) will
|
||||
* not write more than size-1 bytes of text into @buf.
|
||||
*/
|
||||
seq = find_first_fitting_seq(iter->cur_seq, iter->next_seq,
|
||||
size - 1, syslog, time);
|
||||
|
||||
/* move first record forward until length fits into the buffer */
|
||||
seq = dumper->cur_seq;
|
||||
while (l >= size && prb_read_valid_info(prb, seq,
|
||||
&info, &line_count)) {
|
||||
if (r.info->seq >= dumper->next_seq)
|
||||
break;
|
||||
l -= get_record_print_text_size(&info, line_count, syslog, time);
|
||||
seq = r.info->seq + 1;
|
||||
}
|
||||
|
||||
/* last message in next interation */
|
||||
/*
|
||||
* Next kmsg_dump_get_buffer() invocation will dump block of
|
||||
* older records stored right before this one.
|
||||
*/
|
||||
next_seq = seq;
|
||||
|
||||
/* actually read text into the buffer now */
|
||||
l = 0;
|
||||
while (prb_read_valid(prb, seq, &r)) {
|
||||
if (r.info->seq >= dumper->next_seq)
|
||||
prb_rec_init_rd(&r, &info, buf, size);
|
||||
|
||||
len = 0;
|
||||
prb_for_each_record(seq, prb, seq, &r) {
|
||||
if (r.info->seq >= iter->next_seq)
|
||||
break;
|
||||
|
||||
l += record_print_text(&r, syslog, time);
|
||||
len += record_print_text(&r, syslog, time);
|
||||
|
||||
/* adjust record to store to remaining buffer space */
|
||||
prb_rec_init_rd(&r, &info, buf + l, size - l);
|
||||
|
||||
seq = r.info->seq + 1;
|
||||
/* Adjust record to store to remaining buffer space. */
|
||||
prb_rec_init_rd(&r, &info, buf + len, size - len);
|
||||
}
|
||||
|
||||
dumper->next_seq = next_seq;
|
||||
iter->next_seq = next_seq;
|
||||
ret = true;
|
||||
logbuf_unlock_irqrestore(flags);
|
||||
printk_safe_exit_irqrestore(flags);
|
||||
out:
|
||||
if (len)
|
||||
*len = l;
|
||||
if (len_out)
|
||||
*len_out = len;
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(kmsg_dump_get_buffer);
|
||||
|
||||
/**
|
||||
* kmsg_dump_rewind_nolock - reset the iterator (unlocked version)
|
||||
* @dumper: registered kmsg dumper
|
||||
*
|
||||
* Reset the dumper's iterator so that kmsg_dump_get_line() and
|
||||
* kmsg_dump_get_buffer() can be called again and used multiple
|
||||
* times within the same dumper.dump() callback.
|
||||
*
|
||||
* The function is similar to kmsg_dump_rewind(), but grabs no locks.
|
||||
*/
|
||||
void kmsg_dump_rewind_nolock(struct kmsg_dumper *dumper)
|
||||
{
|
||||
dumper->cur_seq = clear_seq;
|
||||
dumper->next_seq = prb_next_seq(prb);
|
||||
}
|
||||
|
||||
/**
|
||||
* kmsg_dump_rewind - reset the iterator
|
||||
* @dumper: registered kmsg dumper
|
||||
* @iter: kmsg dump iterator
|
||||
*
|
||||
* Reset the dumper's iterator so that kmsg_dump_get_line() and
|
||||
* kmsg_dump_get_buffer() can be called again and used multiple
|
||||
* times within the same dumper.dump() callback.
|
||||
*/
|
||||
void kmsg_dump_rewind(struct kmsg_dumper *dumper)
|
||||
void kmsg_dump_rewind(struct kmsg_dump_iter *iter)
|
||||
{
|
||||
unsigned long flags;
|
||||
|
||||
logbuf_lock_irqsave(flags);
|
||||
kmsg_dump_rewind_nolock(dumper);
|
||||
logbuf_unlock_irqrestore(flags);
|
||||
printk_safe_enter_irqsave(flags);
|
||||
iter->cur_seq = latched_seq_read_nolock(&clear_seq);
|
||||
iter->next_seq = prb_next_seq(prb);
|
||||
printk_safe_exit_irqrestore(flags);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(kmsg_dump_rewind);
|
||||
|
||||
|
@ -16,7 +16,7 @@
|
||||
#include "internal.h"
|
||||
|
||||
/*
|
||||
* printk() could not take logbuf_lock in NMI context. Instead,
|
||||
* In NMI and safe mode, printk() avoids taking locks. Instead,
|
||||
* it uses an alternative implementation that temporary stores
|
||||
* the strings into a per-CPU buffer. The content of the buffer
|
||||
* is later flushed into the main ring buffer via IRQ work.
|
||||
@ -267,17 +267,9 @@ void printk_safe_flush(void)
|
||||
void printk_safe_flush_on_panic(void)
|
||||
{
|
||||
/*
|
||||
* Make sure that we could access the main ring buffer.
|
||||
* Make sure that we could access the safe buffers.
|
||||
* Do not risk a double release when more CPUs are up.
|
||||
*/
|
||||
if (raw_spin_is_locked(&logbuf_lock)) {
|
||||
if (num_online_cpus() > 1)
|
||||
return;
|
||||
|
||||
debug_locks_off();
|
||||
raw_spin_lock_init(&logbuf_lock);
|
||||
}
|
||||
|
||||
if (raw_spin_is_locked(&safe_read_lock)) {
|
||||
if (num_online_cpus() > 1)
|
||||
return;
|
||||
@ -319,9 +311,7 @@ void noinstr printk_nmi_exit(void)
|
||||
* reordering.
|
||||
*
|
||||
* It has effect only when called in NMI context. Then printk()
|
||||
* will try to store the messages into the main logbuf directly
|
||||
* and use the per-CPU buffers only as a fallback when the lock
|
||||
* is not available.
|
||||
* will store the messages into the main logbuf directly.
|
||||
*/
|
||||
void printk_nmi_direct_enter(void)
|
||||
{
|
||||
@ -367,7 +357,7 @@ void __printk_safe_exit(void)
|
||||
this_cpu_dec(printk_context);
|
||||
}
|
||||
|
||||
__printf(1, 0) int vprintk_func(const char *fmt, va_list args)
|
||||
asmlinkage int vprintk(const char *fmt, va_list args)
|
||||
{
|
||||
#ifdef CONFIG_KGDB_KDB
|
||||
/* Allow to pass printk() to kdb but avoid a recursion. */
|
||||
@ -376,20 +366,21 @@ __printf(1, 0) int vprintk_func(const char *fmt, va_list args)
|
||||
#endif
|
||||
|
||||
/*
|
||||
* Try to use the main logbuf even in NMI. But avoid calling console
|
||||
* Use the main logbuf even in NMI. But avoid calling console
|
||||
* drivers that might have their own locks.
|
||||
*/
|
||||
if ((this_cpu_read(printk_context) & PRINTK_NMI_DIRECT_CONTEXT_MASK) &&
|
||||
raw_spin_trylock(&logbuf_lock)) {
|
||||
if ((this_cpu_read(printk_context) & PRINTK_NMI_DIRECT_CONTEXT_MASK)) {
|
||||
unsigned long flags;
|
||||
int len;
|
||||
|
||||
printk_safe_enter_irqsave(flags);
|
||||
len = vprintk_store(0, LOGLEVEL_DEFAULT, NULL, fmt, args);
|
||||
raw_spin_unlock(&logbuf_lock);
|
||||
printk_safe_exit_irqrestore(flags);
|
||||
defer_console_output();
|
||||
return len;
|
||||
}
|
||||
|
||||
/* Use extra buffer in NMI when logbuf_lock is taken or in safe mode. */
|
||||
/* Use extra buffer in NMI. */
|
||||
if (this_cpu_read(printk_context) & PRINTK_NMI_CONTEXT_MASK)
|
||||
return vprintk_nmi(fmt, args);
|
||||
|
||||
@ -420,3 +411,4 @@ void __init printk_safe_init(void)
|
||||
/* Flush pending messages that did not have scheduled IRQ works. */
|
||||
printk_safe_flush();
|
||||
}
|
||||
EXPORT_SYMBOL(vprintk);
|
||||
|
@ -577,24 +577,98 @@ netdev_features(void)
|
||||
{
|
||||
}
|
||||
|
||||
struct page_flags_test {
|
||||
int width;
|
||||
int shift;
|
||||
int mask;
|
||||
unsigned long value;
|
||||
const char *fmt;
|
||||
const char *name;
|
||||
};
|
||||
|
||||
static struct page_flags_test pft[] = {
|
||||
{SECTIONS_WIDTH, SECTIONS_PGSHIFT, SECTIONS_MASK,
|
||||
0, "%d", "section"},
|
||||
{NODES_WIDTH, NODES_PGSHIFT, NODES_MASK,
|
||||
0, "%d", "node"},
|
||||
{ZONES_WIDTH, ZONES_PGSHIFT, ZONES_MASK,
|
||||
0, "%d", "zone"},
|
||||
{LAST_CPUPID_WIDTH, LAST_CPUPID_PGSHIFT, LAST_CPUPID_MASK,
|
||||
0, "%#x", "lastcpupid"},
|
||||
{KASAN_TAG_WIDTH, KASAN_TAG_PGSHIFT, KASAN_TAG_MASK,
|
||||
0, "%#x", "kasantag"},
|
||||
};
|
||||
|
||||
static void __init
|
||||
page_flags_test(int section, int node, int zone, int last_cpupid,
|
||||
int kasan_tag, int flags, const char *name, char *cmp_buf)
|
||||
{
|
||||
unsigned long values[] = {section, node, zone, last_cpupid, kasan_tag};
|
||||
unsigned long page_flags = 0;
|
||||
unsigned long size = 0;
|
||||
bool append = false;
|
||||
int i;
|
||||
|
||||
flags &= BIT(NR_PAGEFLAGS) - 1;
|
||||
if (flags) {
|
||||
page_flags |= flags;
|
||||
snprintf(cmp_buf + size, BUF_SIZE - size, "%s", name);
|
||||
size = strlen(cmp_buf);
|
||||
#if SECTIONS_WIDTH || NODES_WIDTH || ZONES_WIDTH || \
|
||||
LAST_CPUPID_WIDTH || KASAN_TAG_WIDTH
|
||||
/* Other information also included in page flags */
|
||||
snprintf(cmp_buf + size, BUF_SIZE - size, "|");
|
||||
size = strlen(cmp_buf);
|
||||
#endif
|
||||
}
|
||||
|
||||
/* Set the test value */
|
||||
for (i = 0; i < ARRAY_SIZE(pft); i++)
|
||||
pft[i].value = values[i];
|
||||
|
||||
for (i = 0; i < ARRAY_SIZE(pft); i++) {
|
||||
if (!pft[i].width)
|
||||
continue;
|
||||
|
||||
if (append) {
|
||||
snprintf(cmp_buf + size, BUF_SIZE - size, "|");
|
||||
size = strlen(cmp_buf);
|
||||
}
|
||||
|
||||
page_flags |= (pft[i].value & pft[i].mask) << pft[i].shift;
|
||||
snprintf(cmp_buf + size, BUF_SIZE - size, "%s=", pft[i].name);
|
||||
size = strlen(cmp_buf);
|
||||
snprintf(cmp_buf + size, BUF_SIZE - size, pft[i].fmt,
|
||||
pft[i].value & pft[i].mask);
|
||||
size = strlen(cmp_buf);
|
||||
append = true;
|
||||
}
|
||||
|
||||
test(cmp_buf, "%pGp", &page_flags);
|
||||
}
|
||||
|
||||
static void __init
|
||||
flags(void)
|
||||
{
|
||||
unsigned long flags;
|
||||
gfp_t gfp;
|
||||
char *cmp_buffer;
|
||||
gfp_t gfp;
|
||||
|
||||
cmp_buffer = kmalloc(BUF_SIZE, GFP_KERNEL);
|
||||
if (!cmp_buffer)
|
||||
return;
|
||||
|
||||
flags = 0;
|
||||
test("", "%pGp", &flags);
|
||||
page_flags_test(0, 0, 0, 0, 0, flags, "", cmp_buffer);
|
||||
|
||||
/* Page flags should filter the zone id */
|
||||
flags = 1UL << NR_PAGEFLAGS;
|
||||
test("", "%pGp", &flags);
|
||||
page_flags_test(0, 0, 0, 0, 0, flags, "", cmp_buffer);
|
||||
|
||||
flags |= 1UL << PG_uptodate | 1UL << PG_dirty | 1UL << PG_lru
|
||||
| 1UL << PG_active | 1UL << PG_swapbacked;
|
||||
test("uptodate|dirty|lru|active|swapbacked", "%pGp", &flags);
|
||||
|
||||
page_flags_test(1, 1, 1, 0x1fffff, 1, flags,
|
||||
"uptodate|dirty|lru|active|swapbacked",
|
||||
cmp_buffer);
|
||||
|
||||
flags = VM_READ | VM_EXEC | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC
|
||||
| VM_DENYWRITE;
|
||||
@ -609,10 +683,6 @@ flags(void)
|
||||
gfp = __GFP_ATOMIC;
|
||||
test("__GFP_ATOMIC", "%pGg", &gfp);
|
||||
|
||||
cmp_buffer = kmalloc(BUF_SIZE, GFP_KERNEL);
|
||||
if (!cmp_buffer)
|
||||
return;
|
||||
|
||||
/* Any flags not translated by the table should remain numeric */
|
||||
gfp = ~__GFP_BITS_MASK;
|
||||
snprintf(cmp_buffer, BUF_SIZE, "%#lx", (unsigned long) gfp);
|
||||
|
@ -1916,6 +1916,66 @@ char *format_flags(char *buf, char *end, unsigned long flags,
|
||||
return buf;
|
||||
}
|
||||
|
||||
struct page_flags_fields {
|
||||
int width;
|
||||
int shift;
|
||||
int mask;
|
||||
const struct printf_spec *spec;
|
||||
const char *name;
|
||||
};
|
||||
|
||||
static const struct page_flags_fields pff[] = {
|
||||
{SECTIONS_WIDTH, SECTIONS_PGSHIFT, SECTIONS_MASK,
|
||||
&default_dec_spec, "section"},
|
||||
{NODES_WIDTH, NODES_PGSHIFT, NODES_MASK,
|
||||
&default_dec_spec, "node"},
|
||||
{ZONES_WIDTH, ZONES_PGSHIFT, ZONES_MASK,
|
||||
&default_dec_spec, "zone"},
|
||||
{LAST_CPUPID_WIDTH, LAST_CPUPID_PGSHIFT, LAST_CPUPID_MASK,
|
||||
&default_flag_spec, "lastcpupid"},
|
||||
{KASAN_TAG_WIDTH, KASAN_TAG_PGSHIFT, KASAN_TAG_MASK,
|
||||
&default_flag_spec, "kasantag"},
|
||||
};
|
||||
|
||||
static
|
||||
char *format_page_flags(char *buf, char *end, unsigned long flags)
|
||||
{
|
||||
unsigned long main_flags = flags & (BIT(NR_PAGEFLAGS) - 1);
|
||||
bool append = false;
|
||||
int i;
|
||||
|
||||
/* Page flags from the main area. */
|
||||
if (main_flags) {
|
||||
buf = format_flags(buf, end, main_flags, pageflag_names);
|
||||
append = true;
|
||||
}
|
||||
|
||||
/* Page flags from the fields area */
|
||||
for (i = 0; i < ARRAY_SIZE(pff); i++) {
|
||||
/* Skip undefined fields. */
|
||||
if (!pff[i].width)
|
||||
continue;
|
||||
|
||||
/* Format: Flag Name + '=' (equals sign) + Number + '|' (separator) */
|
||||
if (append) {
|
||||
if (buf < end)
|
||||
*buf = '|';
|
||||
buf++;
|
||||
}
|
||||
|
||||
buf = string(buf, end, pff[i].name, default_str_spec);
|
||||
if (buf < end)
|
||||
*buf = '=';
|
||||
buf++;
|
||||
buf = number(buf, end, (flags >> pff[i].shift) & pff[i].mask,
|
||||
*pff[i].spec);
|
||||
|
||||
append = true;
|
||||
}
|
||||
|
||||
return buf;
|
||||
}
|
||||
|
||||
static noinline_for_stack
|
||||
char *flags_string(char *buf, char *end, void *flags_ptr,
|
||||
struct printf_spec spec, const char *fmt)
|
||||
@ -1928,11 +1988,7 @@ char *flags_string(char *buf, char *end, void *flags_ptr,
|
||||
|
||||
switch (fmt[1]) {
|
||||
case 'p':
|
||||
flags = *(unsigned long *)flags_ptr;
|
||||
/* Remove zone id */
|
||||
flags &= (1UL << NR_PAGEFLAGS) - 1;
|
||||
names = pageflag_names;
|
||||
break;
|
||||
return format_page_flags(buf, end, *(unsigned long *)flags_ptr);
|
||||
case 'v':
|
||||
flags = *(unsigned long *)flags_ptr;
|
||||
names = vmaflag_names;
|
||||
@ -2096,6 +2152,9 @@ EXPORT_SYMBOL_GPL(no_hash_pointers);
|
||||
|
||||
static int __init no_hash_pointers_enable(char *str)
|
||||
{
|
||||
if (no_hash_pointers)
|
||||
return 0;
|
||||
|
||||
no_hash_pointers = true;
|
||||
|
||||
pr_warn("**********************************************************\n");
|
||||
@ -2186,7 +2245,9 @@ early_param("no_hash_pointers", no_hash_pointers_enable);
|
||||
* Implements a "recursive vsnprintf".
|
||||
* Do not use this feature without some mechanism to verify the
|
||||
* correctness of the format string and va_list arguments.
|
||||
* - 'K' For a kernel pointer that should be hidden from unprivileged users
|
||||
* - 'K' For a kernel pointer that should be hidden from unprivileged users.
|
||||
* Use only for procfs, sysfs and similar files, not printk(); please
|
||||
* read the documentation (path below) first.
|
||||
* - 'NF' For a netdev_features_t
|
||||
* - 'h[CDN]' For a variable-length buffer, it prints it as a hex string with
|
||||
* a certain separator (' ' by default):
|
||||
@ -2225,7 +2286,8 @@ early_param("no_hash_pointers", no_hash_pointers_enable);
|
||||
* Without an option prints the full name of the node
|
||||
* f full name
|
||||
* P node name, including a possible unit address
|
||||
* - 'x' For printing the address. Equivalent to "%lx".
|
||||
* - 'x' For printing the address unmodified. Equivalent to "%lx".
|
||||
* Please read the documentation (path below) before using!
|
||||
* - '[ku]s' For a BPF/tracing related format specifier, e.g. used out of
|
||||
* bpf_trace_printk() where [ku] prefix specifies either kernel (k)
|
||||
* or user (u) memory to probe, and:
|
||||
@ -3135,8 +3197,6 @@ int bstr_printf(char *buf, size_t size, const char *fmt, const u32 *bin_buf)
|
||||
switch (*fmt) {
|
||||
case 'S':
|
||||
case 's':
|
||||
case 'F':
|
||||
case 'f':
|
||||
case 'x':
|
||||
case 'K':
|
||||
case 'e':
|
||||
|
13
mm/slub.c
13
mm/slub.c
@ -624,7 +624,7 @@ static void print_track(const char *s, struct track *t, unsigned long pr_time)
|
||||
if (!t->addr)
|
||||
return;
|
||||
|
||||
pr_err("INFO: %s in %pS age=%lu cpu=%u pid=%d\n",
|
||||
pr_err("%s in %pS age=%lu cpu=%u pid=%d\n",
|
||||
s, (void *)t->addr, pr_time - t->when, t->cpu, t->pid);
|
||||
#ifdef CONFIG_STACKTRACE
|
||||
{
|
||||
@ -650,8 +650,9 @@ void print_tracking(struct kmem_cache *s, void *object)
|
||||
|
||||
static void print_page_info(struct page *page)
|
||||
{
|
||||
pr_err("INFO: Slab 0x%p objects=%u used=%u fp=0x%p flags=0x%04lx\n",
|
||||
page, page->objects, page->inuse, page->freelist, page->flags);
|
||||
pr_err("Slab 0x%p objects=%u used=%u fp=0x%p flags=%#lx(%pGp)\n",
|
||||
page, page->objects, page->inuse, page->freelist,
|
||||
page->flags, &page->flags);
|
||||
|
||||
}
|
||||
|
||||
@ -706,7 +707,7 @@ static void print_trailer(struct kmem_cache *s, struct page *page, u8 *p)
|
||||
|
||||
print_page_info(page);
|
||||
|
||||
pr_err("INFO: Object 0x%p @offset=%tu fp=0x%p\n\n",
|
||||
pr_err("Object 0x%p @offset=%tu fp=0x%p\n\n",
|
||||
p, p - addr, get_freepointer(s, p));
|
||||
|
||||
if (s->flags & SLAB_RED_ZONE)
|
||||
@ -799,7 +800,7 @@ static int check_bytes_and_report(struct kmem_cache *s, struct page *page,
|
||||
end--;
|
||||
|
||||
slab_bug(s, "%s overwritten", what);
|
||||
pr_err("INFO: 0x%p-0x%p @offset=%tu. First byte 0x%x instead of 0x%x\n",
|
||||
pr_err("0x%p-0x%p @offset=%tu. First byte 0x%x instead of 0x%x\n",
|
||||
fault, end - 1, fault - addr,
|
||||
fault[0], value);
|
||||
print_trailer(s, page, object);
|
||||
@ -3898,7 +3899,7 @@ static void list_slab_objects(struct kmem_cache *s, struct page *page,
|
||||
for_each_object(p, s, addr, page->objects) {
|
||||
|
||||
if (!test_bit(__obj_to_index(s, addr, p), map)) {
|
||||
pr_err("INFO: Object 0x%p @offset=%tu\n", p, p - addr);
|
||||
pr_err("Object 0x%p @offset=%tu\n", p, p - addr);
|
||||
print_tracking(s, p);
|
||||
}
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user