mirror of
https://github.com/torvalds/linux.git
synced 2024-12-26 12:52:30 +00:00
tracing updates for 6.4:
- User events are finally ready! After lots of collaboration between various parties, we finally locked down on a stable interface for user events that can also work with user space only tracing. This is implemented by telling the kernel (or user space library, but that part is user space only and not part of this patch set), where the variable is that the application uses to know if something is listening to the trace. There's also an interface to tell the kernel about these events, which will show up in the /sys/kernel/tracing/events/user_events/ directory, where it can be enabled. When it's enabled, the kernel will update the variable, to tell the application to start writing to the kernel. See https://lwn.net/Articles/927595/ - Cleaned up the direct trampolines code to simplify arm64 addition of direct trampolines. Direct trampolines use the ftrace interface but instead of jumping to the ftrace trampoline, applications (mostly BPF) can register their own trampoline for performance reasons. - Some updates to the fprobe infrastructure. fprobes are more efficient than kprobes, as it does not need to save all the registers that kprobes on ftrace do. More work needs to be done before the fprobes will be exposed as dynamic events. - More updates to references to the obsolete path of /sys/kernel/debug/tracing for the new /sys/kernel/tracing path. - Add a seq_buf_do_printk() helper to seq_bufs, to print a large buffer line by line instead of all at once. There's users in production kernels that have a large data dump that originally used printk() directly, but the data dump was larger than what printk() allowed as a single print. Using seq_buf() to do the printing fixes that. - Add /sys/kernel/tracing/touched_functions that shows all functions that was every traced by ftrace or a direct trampoline. This is used for debugging issues where a traced function could have caused a crash by a bpf program or live patching. - Add a "fields" option that is similar to "raw" but outputs the fields of the events. It's easier to read by humans. - Some minor fixes and clean ups. -----BEGIN PGP SIGNATURE----- iIoEABYIADIWIQRRSw7ePDh/lE+zeZMp5XQQmuv6qgUCZEr36xQccm9zdGVkdEBn b29kbWlzLm9yZwAKCRAp5XQQmuv6quZHAQCzuqnn2S8DsPd3Sy1vKIYaj0uajW5D Kz1oUJH4F0H7kgEA8XwXkdtfKpOXWc/ZH4LWfL7Orx2wJZJQMV9dVqEPDAE= =w0Z1 -----END PGP SIGNATURE----- Merge tag 'trace-v6.4' of git://git.kernel.org/pub/scm/linux/kernel/git/trace/linux-trace Pull tracing updates from Steven Rostedt: - User events are finally ready! After lots of collaboration between various parties, we finally locked down on a stable interface for user events that can also work with user space only tracing. This is implemented by telling the kernel (or user space library, but that part is user space only and not part of this patch set), where the variable is that the application uses to know if something is listening to the trace. There's also an interface to tell the kernel about these events, which will show up in the /sys/kernel/tracing/events/user_events/ directory, where it can be enabled. When it's enabled, the kernel will update the variable, to tell the application to start writing to the kernel. See https://lwn.net/Articles/927595/ - Cleaned up the direct trampolines code to simplify arm64 addition of direct trampolines. Direct trampolines use the ftrace interface but instead of jumping to the ftrace trampoline, applications (mostly BPF) can register their own trampoline for performance reasons. - Some updates to the fprobe infrastructure. fprobes are more efficient than kprobes, as it does not need to save all the registers that kprobes on ftrace do. More work needs to be done before the fprobes will be exposed as dynamic events. - More updates to references to the obsolete path of /sys/kernel/debug/tracing for the new /sys/kernel/tracing path. - Add a seq_buf_do_printk() helper to seq_bufs, to print a large buffer line by line instead of all at once. There are users in production kernels that have a large data dump that originally used printk() directly, but the data dump was larger than what printk() allowed as a single print. Using seq_buf() to do the printing fixes that. - Add /sys/kernel/tracing/touched_functions that shows all functions that was every traced by ftrace or a direct trampoline. This is used for debugging issues where a traced function could have caused a crash by a bpf program or live patching. - Add a "fields" option that is similar to "raw" but outputs the fields of the events. It's easier to read by humans. - Some minor fixes and clean ups. * tag 'trace-v6.4' of git://git.kernel.org/pub/scm/linux/kernel/git/trace/linux-trace: (41 commits) ring-buffer: Sync IRQ works before buffer destruction tracing: Add missing spaces in trace_print_hex_seq() ring-buffer: Ensure proper resetting of atomic variables in ring_buffer_reset_online_cpus recordmcount: Fix memory leaks in the uwrite function tracing/user_events: Limit max fault-in attempts tracing/user_events: Prevent same address and bit per process tracing/user_events: Ensure bit is cleared on unregister tracing/user_events: Ensure write index cannot be negative seq_buf: Add seq_buf_do_printk() helper tracing: Fix print_fields() for __dyn_loc/__rel_loc tracing/user_events: Set event filter_type from type ring-buffer: Clearly check null ptr returned by rb_set_head_page() tracing: Unbreak user events tracing/user_events: Use print_format_fields() for trace output tracing/user_events: Align structs with tabs for readability tracing/user_events: Limit global user_event count tracing/user_events: Charge event allocs to cgroups tracing/user_events: Update documentation for ABI tracing/user_events: Use write ABI in example tracing/user_events: Add ABI self-test ...
This commit is contained in:
commit
d579c468d7
@ -87,14 +87,16 @@ returns as same as unregister_ftrace_function().
|
||||
The fprobe entry/exit handler
|
||||
=============================
|
||||
|
||||
The prototype of the entry/exit callback function is as follows:
|
||||
The prototype of the entry/exit callback function are as follows:
|
||||
|
||||
.. code-block:: c
|
||||
|
||||
void callback_func(struct fprobe *fp, unsigned long entry_ip, struct pt_regs *regs);
|
||||
int entry_callback(struct fprobe *fp, unsigned long entry_ip, struct pt_regs *regs, void *entry_data);
|
||||
|
||||
Note that both entry and exit callbacks have same ptototype. The @entry_ip is
|
||||
saved at function entry and passed to exit handler.
|
||||
void exit_callback(struct fprobe *fp, unsigned long entry_ip, struct pt_regs *regs, void *entry_data);
|
||||
|
||||
Note that the @entry_ip is saved at function entry and passed to exit handler.
|
||||
If the entry callback function returns !0, the corresponding exit callback will be cancelled.
|
||||
|
||||
@fp
|
||||
This is the address of `fprobe` data structure related to this handler.
|
||||
@ -113,6 +115,12 @@ saved at function entry and passed to exit handler.
|
||||
to use @entry_ip. On the other hand, in the exit_handler, the instruction
|
||||
pointer of @regs is set to the currect return address.
|
||||
|
||||
@entry_data
|
||||
This is a local storage to share the data between entry and exit handlers.
|
||||
This storage is NULL by default. If the user specify `exit_handler` field
|
||||
and `entry_data_size` field when registering the fprobe, the storage is
|
||||
allocated and passed to both `entry_handler` and `exit_handler`.
|
||||
|
||||
Share the callbacks with kprobes
|
||||
================================
|
||||
|
||||
|
@ -1027,6 +1027,7 @@ To see what is available, simply cat the file::
|
||||
nohex
|
||||
nobin
|
||||
noblock
|
||||
nofields
|
||||
trace_printk
|
||||
annotate
|
||||
nouserstacktrace
|
||||
@ -1110,6 +1111,11 @@ Here are the available options:
|
||||
block
|
||||
When set, reading trace_pipe will not block when polled.
|
||||
|
||||
fields
|
||||
Print the fields as described by their types. This is a better
|
||||
option than using hex, bin or raw, as it gives a better parsing
|
||||
of the content of the event.
|
||||
|
||||
trace_printk
|
||||
Can disable trace_printk() from writing into the buffer.
|
||||
|
||||
|
@ -20,11 +20,10 @@ dynamic_events is the same as the ioctl with the u: prefix applied.
|
||||
|
||||
Typically programs will register a set of events that they wish to expose to
|
||||
tools that can read trace_events (such as ftrace and perf). The registration
|
||||
process gives back two ints to the program for each event. The first int is
|
||||
the status bit. This describes which bit in little-endian format in the
|
||||
/sys/kernel/tracing/user_events_status file represents this event. The
|
||||
second int is the write index which describes the data when a write() or
|
||||
writev() is called on the /sys/kernel/tracing/user_events_data file.
|
||||
process tells the kernel which address and bit to reflect if any tool has
|
||||
enabled the event and data should be written. The registration will give back
|
||||
a write index which describes the data when a write() or writev() is called
|
||||
on the /sys/kernel/tracing/user_events_data file.
|
||||
|
||||
The structures referenced in this document are contained within the
|
||||
/include/uapi/linux/user_events.h file in the source tree.
|
||||
@ -41,23 +40,64 @@ DIAG_IOCSREG.
|
||||
This command takes a packed struct user_reg as an argument::
|
||||
|
||||
struct user_reg {
|
||||
u32 size;
|
||||
u64 name_args;
|
||||
u32 status_bit;
|
||||
u32 write_index;
|
||||
};
|
||||
/* Input: Size of the user_reg structure being used */
|
||||
__u32 size;
|
||||
|
||||
The struct user_reg requires two inputs, the first is the size of the structure
|
||||
to ensure forward and backward compatibility. The second is the command string
|
||||
to issue for registering. Upon success two outputs are set, the status bit
|
||||
and the write index.
|
||||
/* Input: Bit in enable address to use */
|
||||
__u8 enable_bit;
|
||||
|
||||
/* Input: Enable size in bytes at address */
|
||||
__u8 enable_size;
|
||||
|
||||
/* Input: Flags for future use, set to 0 */
|
||||
__u16 flags;
|
||||
|
||||
/* Input: Address to update when enabled */
|
||||
__u64 enable_addr;
|
||||
|
||||
/* Input: Pointer to string with event name, description and flags */
|
||||
__u64 name_args;
|
||||
|
||||
/* Output: Index of the event to use when writing data */
|
||||
__u32 write_index;
|
||||
} __attribute__((__packed__));
|
||||
|
||||
The struct user_reg requires all the above inputs to be set appropriately.
|
||||
|
||||
+ size: This must be set to sizeof(struct user_reg).
|
||||
|
||||
+ enable_bit: The bit to reflect the event status at the address specified by
|
||||
enable_addr.
|
||||
|
||||
+ enable_size: The size of the value specified by enable_addr.
|
||||
This must be 4 (32-bit) or 8 (64-bit). 64-bit values are only allowed to be
|
||||
used on 64-bit kernels, however, 32-bit can be used on all kernels.
|
||||
|
||||
+ flags: The flags to use, if any. For the initial version this must be 0.
|
||||
Callers should first attempt to use flags and retry without flags to ensure
|
||||
support for lower versions of the kernel. If a flag is not supported -EINVAL
|
||||
is returned.
|
||||
|
||||
+ enable_addr: The address of the value to use to reflect event status. This
|
||||
must be naturally aligned and write accessible within the user program.
|
||||
|
||||
+ name_args: The name and arguments to describe the event, see command format
|
||||
for details.
|
||||
|
||||
Upon successful registration the following is set.
|
||||
|
||||
+ write_index: The index to use for this file descriptor that represents this
|
||||
event when writing out data. The index is unique to this instance of the file
|
||||
descriptor that was used for the registration. See writing data for details.
|
||||
|
||||
User based events show up under tracefs like any other event under the
|
||||
subsystem named "user_events". This means tools that wish to attach to the
|
||||
events need to use /sys/kernel/tracing/events/user_events/[name]/enable
|
||||
or perf record -e user_events:[name] when attaching/recording.
|
||||
|
||||
**NOTE:** *The write_index returned is only valid for the FD that was used*
|
||||
**NOTE:** The event subsystem name by default is "user_events". Callers should
|
||||
not assume it will always be "user_events". Operators reserve the right in the
|
||||
future to change the subsystem name per-process to accomodate event isolation.
|
||||
|
||||
Command Format
|
||||
^^^^^^^^^^^^^^
|
||||
@ -94,7 +134,7 @@ Would be represented by the following field::
|
||||
struct mytype myname 20
|
||||
|
||||
Deleting
|
||||
-----------
|
||||
--------
|
||||
Deleting an event from within a user process is done via ioctl() out to the
|
||||
/sys/kernel/tracing/user_events_data file. The command to issue is
|
||||
DIAG_IOCSDEL.
|
||||
@ -104,92 +144,79 @@ its name. Delete will only succeed if there are no references left to the
|
||||
event (in both user and kernel space). User programs should use a separate file
|
||||
to request deletes than the one used for registration due to this.
|
||||
|
||||
Unregistering
|
||||
-------------
|
||||
If after registering an event it is no longer wanted to be updated then it can
|
||||
be disabled via ioctl() out to the /sys/kernel/tracing/user_events_data file.
|
||||
The command to issue is DIAG_IOCSUNREG. This is different than deleting, where
|
||||
deleting actually removes the event from the system. Unregistering simply tells
|
||||
the kernel your process is no longer interested in updates to the event.
|
||||
|
||||
This command takes a packed struct user_unreg as an argument::
|
||||
|
||||
struct user_unreg {
|
||||
/* Input: Size of the user_unreg structure being used */
|
||||
__u32 size;
|
||||
|
||||
/* Input: Bit to unregister */
|
||||
__u8 disable_bit;
|
||||
|
||||
/* Input: Reserved, set to 0 */
|
||||
__u8 __reserved;
|
||||
|
||||
/* Input: Reserved, set to 0 */
|
||||
__u16 __reserved2;
|
||||
|
||||
/* Input: Address to unregister */
|
||||
__u64 disable_addr;
|
||||
} __attribute__((__packed__));
|
||||
|
||||
The struct user_unreg requires all the above inputs to be set appropriately.
|
||||
|
||||
+ size: This must be set to sizeof(struct user_unreg).
|
||||
|
||||
+ disable_bit: This must be set to the bit to disable (same bit that was
|
||||
previously registered via enable_bit).
|
||||
|
||||
+ disable_addr: This must be set to the address to disable (same address that was
|
||||
previously registered via enable_addr).
|
||||
|
||||
**NOTE:** Events are automatically unregistered when execve() is invoked. During
|
||||
fork() the registered events will be retained and must be unregistered manually
|
||||
in each process if wanted.
|
||||
|
||||
Status
|
||||
------
|
||||
When tools attach/record user based events the status of the event is updated
|
||||
in realtime. This allows user programs to only incur the cost of the write() or
|
||||
writev() calls when something is actively attached to the event.
|
||||
|
||||
User programs call mmap() on /sys/kernel/tracing/user_events_status to
|
||||
check the status for each event that is registered. The bit to check in the
|
||||
file is given back after the register ioctl() via user_reg.status_bit. The bit
|
||||
is always in little-endian format. Programs can check if the bit is set either
|
||||
using a byte-wise index with a mask or a long-wise index with a little-endian
|
||||
mask.
|
||||
|
||||
Currently the size of user_events_status is a single page, however, custom
|
||||
kernel configurations can change this size to allow more user based events. In
|
||||
all cases the size of the file is a multiple of a page size.
|
||||
|
||||
For example, if the register ioctl() gives back a status_bit of 3 you would
|
||||
check byte 0 (3 / 8) of the returned mmap data and then AND the result with 8
|
||||
(1 << (3 % 8)) to see if anything is attached to that event.
|
||||
|
||||
A byte-wise index check is performed as follows::
|
||||
|
||||
int index, mask;
|
||||
char *status_page;
|
||||
|
||||
index = status_bit / 8;
|
||||
mask = 1 << (status_bit % 8);
|
||||
|
||||
...
|
||||
|
||||
if (status_page[index] & mask) {
|
||||
/* Enabled */
|
||||
}
|
||||
|
||||
A long-wise index check is performed as follows::
|
||||
|
||||
#include <asm/bitsperlong.h>
|
||||
#include <endian.h>
|
||||
|
||||
#if __BITS_PER_LONG == 64
|
||||
#define endian_swap(x) htole64(x)
|
||||
#else
|
||||
#define endian_swap(x) htole32(x)
|
||||
#endif
|
||||
|
||||
long index, mask, *status_page;
|
||||
|
||||
index = status_bit / __BITS_PER_LONG;
|
||||
mask = 1L << (status_bit % __BITS_PER_LONG);
|
||||
mask = endian_swap(mask);
|
||||
|
||||
...
|
||||
|
||||
if (status_page[index] & mask) {
|
||||
/* Enabled */
|
||||
}
|
||||
The kernel will update the specified bit that was registered for the event as
|
||||
tools attach/detach from the event. User programs simply check if the bit is set
|
||||
to see if something is attached or not.
|
||||
|
||||
Administrators can easily check the status of all registered events by reading
|
||||
the user_events_status file directly via a terminal. The output is as follows::
|
||||
|
||||
Byte:Name [# Comments]
|
||||
Name [# Comments]
|
||||
...
|
||||
|
||||
Active: ActiveCount
|
||||
Busy: BusyCount
|
||||
Max: MaxCount
|
||||
|
||||
For example, on a system that has a single event the output looks like this::
|
||||
|
||||
1:test
|
||||
test
|
||||
|
||||
Active: 1
|
||||
Busy: 0
|
||||
Max: 32768
|
||||
|
||||
If a user enables the user event via ftrace, the output would change to this::
|
||||
|
||||
1:test # Used by ftrace
|
||||
test # Used by ftrace
|
||||
|
||||
Active: 1
|
||||
Busy: 1
|
||||
Max: 32768
|
||||
|
||||
**NOTE:** *A status bit of 0 will never be returned. This allows user programs
|
||||
to have a bit that can be used on error cases.*
|
||||
|
||||
Writing Data
|
||||
------------
|
||||
@ -217,7 +244,7 @@ For example, if I have a struct like this::
|
||||
int src;
|
||||
int dst;
|
||||
int flags;
|
||||
};
|
||||
} __attribute__((__packed__));
|
||||
|
||||
It's advised for user programs to do the following::
|
||||
|
||||
|
@ -65,6 +65,7 @@
|
||||
#include <linux/syscall_user_dispatch.h>
|
||||
#include <linux/coredump.h>
|
||||
#include <linux/time_namespace.h>
|
||||
#include <linux/user_events.h>
|
||||
|
||||
#include <linux/uaccess.h>
|
||||
#include <asm/mmu_context.h>
|
||||
@ -1859,6 +1860,7 @@ static int bprm_execve(struct linux_binprm *bprm,
|
||||
current->fs->in_exec = 0;
|
||||
current->in_execve = 0;
|
||||
rseq_execve(current);
|
||||
user_events_execve(current);
|
||||
acct_update_integrals(current);
|
||||
task_numa_free(current, false);
|
||||
return retval;
|
||||
|
@ -13,6 +13,8 @@
|
||||
* @nmissed: The counter for missing events.
|
||||
* @flags: The status flag.
|
||||
* @rethook: The rethook data structure. (internal data)
|
||||
* @entry_data_size: The private data storage size.
|
||||
* @nr_maxactive: The max number of active functions.
|
||||
* @entry_handler: The callback function for function entry.
|
||||
* @exit_handler: The callback function for function exit.
|
||||
*/
|
||||
@ -29,9 +31,13 @@ struct fprobe {
|
||||
unsigned long nmissed;
|
||||
unsigned int flags;
|
||||
struct rethook *rethook;
|
||||
size_t entry_data_size;
|
||||
int nr_maxactive;
|
||||
|
||||
void (*entry_handler)(struct fprobe *fp, unsigned long entry_ip, struct pt_regs *regs);
|
||||
void (*exit_handler)(struct fprobe *fp, unsigned long entry_ip, struct pt_regs *regs);
|
||||
int (*entry_handler)(struct fprobe *fp, unsigned long entry_ip,
|
||||
struct pt_regs *regs, void *entry_data);
|
||||
void (*exit_handler)(struct fprobe *fp, unsigned long entry_ip,
|
||||
struct pt_regs *regs, void *entry_data);
|
||||
};
|
||||
|
||||
/* This fprobe is soft-disabled. */
|
||||
|
@ -548,6 +548,7 @@ bool is_ftrace_trampoline(unsigned long addr);
|
||||
* DIRECT - there is a direct function to call
|
||||
* CALL_OPS - the record can use callsite-specific ops
|
||||
* CALL_OPS_EN - the function is set up to use callsite-specific ops
|
||||
* TOUCHED - A callback was added since boot up
|
||||
*
|
||||
* When a new ftrace_ops is registered and wants a function to save
|
||||
* pt_regs, the rec->flags REGS is set. When the function has been
|
||||
@ -567,9 +568,10 @@ enum {
|
||||
FTRACE_FL_DIRECT_EN = (1UL << 23),
|
||||
FTRACE_FL_CALL_OPS = (1UL << 22),
|
||||
FTRACE_FL_CALL_OPS_EN = (1UL << 21),
|
||||
FTRACE_FL_TOUCHED = (1UL << 20),
|
||||
};
|
||||
|
||||
#define FTRACE_REF_MAX_SHIFT 21
|
||||
#define FTRACE_REF_MAX_SHIFT 20
|
||||
#define FTRACE_REF_MAX ((1UL << FTRACE_REF_MAX_SHIFT) - 1)
|
||||
|
||||
#define ftrace_rec_count(rec) ((rec)->flags & FTRACE_REF_MAX)
|
||||
@ -628,6 +630,7 @@ enum {
|
||||
FTRACE_ITER_PROBE = (1 << 4),
|
||||
FTRACE_ITER_MOD = (1 << 5),
|
||||
FTRACE_ITER_ENABLED = (1 << 6),
|
||||
FTRACE_ITER_TOUCHED = (1 << 7),
|
||||
};
|
||||
|
||||
void arch_ftrace_update_code(int command);
|
||||
|
@ -70,6 +70,7 @@ struct sighand_struct;
|
||||
struct signal_struct;
|
||||
struct task_delay_info;
|
||||
struct task_group;
|
||||
struct user_event_mm;
|
||||
|
||||
/*
|
||||
* Task state bitmask. NOTE! These bits are also
|
||||
@ -1529,6 +1530,10 @@ struct task_struct {
|
||||
union rv_task_monitor rv[RV_PER_TASK_MONITORS];
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_USER_EVENTS
|
||||
struct user_event_mm *user_event_mm;
|
||||
#endif
|
||||
|
||||
/*
|
||||
* New fields for task_struct should be added above here, so that
|
||||
* they are included in the randomized portion of task_struct.
|
||||
|
@ -159,4 +159,6 @@ extern int
|
||||
seq_buf_bprintf(struct seq_buf *s, const char *fmt, const u32 *binary);
|
||||
#endif
|
||||
|
||||
void seq_buf_do_printk(struct seq_buf *s, const char *lvl);
|
||||
|
||||
#endif /* _LINUX_SEQ_BUF_H */
|
||||
|
@ -1,54 +1,83 @@
|
||||
/* SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note */
|
||||
/* SPDX-License-Identifier: GPL-2.0-only */
|
||||
/*
|
||||
* Copyright (c) 2021, Microsoft Corporation.
|
||||
* Copyright (c) 2022, Microsoft Corporation.
|
||||
*
|
||||
* Authors:
|
||||
* Beau Belgrave <beaub@linux.microsoft.com>
|
||||
*/
|
||||
#ifndef _UAPI_LINUX_USER_EVENTS_H
|
||||
#define _UAPI_LINUX_USER_EVENTS_H
|
||||
|
||||
#include <linux/types.h>
|
||||
#include <linux/ioctl.h>
|
||||
#ifndef _LINUX_USER_EVENTS_H
|
||||
#define _LINUX_USER_EVENTS_H
|
||||
|
||||
#ifdef __KERNEL__
|
||||
#include <linux/uio.h>
|
||||
#include <linux/list.h>
|
||||
#include <linux/refcount.h>
|
||||
#include <linux/mm_types.h>
|
||||
#include <linux/workqueue.h>
|
||||
#include <uapi/linux/user_events.h>
|
||||
|
||||
#ifdef CONFIG_USER_EVENTS
|
||||
struct user_event_mm {
|
||||
struct list_head link;
|
||||
struct list_head enablers;
|
||||
struct mm_struct *mm;
|
||||
struct user_event_mm *next;
|
||||
refcount_t refcnt;
|
||||
refcount_t tasks;
|
||||
struct rcu_work put_rwork;
|
||||
};
|
||||
|
||||
extern void user_event_mm_dup(struct task_struct *t,
|
||||
struct user_event_mm *old_mm);
|
||||
|
||||
extern void user_event_mm_remove(struct task_struct *t);
|
||||
|
||||
static inline void user_events_fork(struct task_struct *t,
|
||||
unsigned long clone_flags)
|
||||
{
|
||||
struct user_event_mm *old_mm;
|
||||
|
||||
if (!t || !current->user_event_mm)
|
||||
return;
|
||||
|
||||
old_mm = current->user_event_mm;
|
||||
|
||||
if (clone_flags & CLONE_VM) {
|
||||
t->user_event_mm = old_mm;
|
||||
refcount_inc(&old_mm->tasks);
|
||||
return;
|
||||
}
|
||||
|
||||
user_event_mm_dup(t, old_mm);
|
||||
}
|
||||
|
||||
static inline void user_events_execve(struct task_struct *t)
|
||||
{
|
||||
if (!t || !t->user_event_mm)
|
||||
return;
|
||||
|
||||
user_event_mm_remove(t);
|
||||
}
|
||||
|
||||
static inline void user_events_exit(struct task_struct *t)
|
||||
{
|
||||
if (!t || !t->user_event_mm)
|
||||
return;
|
||||
|
||||
user_event_mm_remove(t);
|
||||
}
|
||||
#else
|
||||
#include <sys/uio.h>
|
||||
#endif
|
||||
static inline void user_events_fork(struct task_struct *t,
|
||||
unsigned long clone_flags)
|
||||
{
|
||||
}
|
||||
|
||||
#define USER_EVENTS_SYSTEM "user_events"
|
||||
#define USER_EVENTS_PREFIX "u:"
|
||||
static inline void user_events_execve(struct task_struct *t)
|
||||
{
|
||||
}
|
||||
|
||||
/* Create dynamic location entry within a 32-bit value */
|
||||
#define DYN_LOC(offset, size) ((size) << 16 | (offset))
|
||||
static inline void user_events_exit(struct task_struct *t)
|
||||
{
|
||||
}
|
||||
#endif /* CONFIG_USER_EVENTS */
|
||||
|
||||
/*
|
||||
* Describes an event registration and stores the results of the registration.
|
||||
* This structure is passed to the DIAG_IOCSREG ioctl, callers at a minimum
|
||||
* must set the size and name_args before invocation.
|
||||
*/
|
||||
struct user_reg {
|
||||
|
||||
/* Input: Size of the user_reg structure being used */
|
||||
__u32 size;
|
||||
|
||||
/* Input: Pointer to string with event name, description and flags */
|
||||
__u64 name_args;
|
||||
|
||||
/* Output: Bitwise index of the event within the status page */
|
||||
__u32 status_bit;
|
||||
|
||||
/* Output: Index of the event to use when writing data */
|
||||
__u32 write_index;
|
||||
} __attribute__((__packed__));
|
||||
|
||||
#define DIAG_IOC_MAGIC '*'
|
||||
|
||||
/* Requests to register a user_event */
|
||||
#define DIAG_IOCSREG _IOWR(DIAG_IOC_MAGIC, 0, struct user_reg*)
|
||||
|
||||
/* Requests to delete a user_event */
|
||||
#define DIAG_IOCSDEL _IOW(DIAG_IOC_MAGIC, 1, char*)
|
||||
|
||||
#endif /* _UAPI_LINUX_USER_EVENTS_H */
|
||||
#endif /* _LINUX_USER_EVENTS_H */
|
||||
|
81
include/uapi/linux/user_events.h
Normal file
81
include/uapi/linux/user_events.h
Normal file
@ -0,0 +1,81 @@
|
||||
/* SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note */
|
||||
/*
|
||||
* Copyright (c) 2021-2022, Microsoft Corporation.
|
||||
*
|
||||
* Authors:
|
||||
* Beau Belgrave <beaub@linux.microsoft.com>
|
||||
*/
|
||||
#ifndef _UAPI_LINUX_USER_EVENTS_H
|
||||
#define _UAPI_LINUX_USER_EVENTS_H
|
||||
|
||||
#include <linux/types.h>
|
||||
#include <linux/ioctl.h>
|
||||
|
||||
#define USER_EVENTS_SYSTEM "user_events"
|
||||
#define USER_EVENTS_PREFIX "u:"
|
||||
|
||||
/* Create dynamic location entry within a 32-bit value */
|
||||
#define DYN_LOC(offset, size) ((size) << 16 | (offset))
|
||||
|
||||
/*
|
||||
* Describes an event registration and stores the results of the registration.
|
||||
* This structure is passed to the DIAG_IOCSREG ioctl, callers at a minimum
|
||||
* must set the size and name_args before invocation.
|
||||
*/
|
||||
struct user_reg {
|
||||
|
||||
/* Input: Size of the user_reg structure being used */
|
||||
__u32 size;
|
||||
|
||||
/* Input: Bit in enable address to use */
|
||||
__u8 enable_bit;
|
||||
|
||||
/* Input: Enable size in bytes at address */
|
||||
__u8 enable_size;
|
||||
|
||||
/* Input: Flags for future use, set to 0 */
|
||||
__u16 flags;
|
||||
|
||||
/* Input: Address to update when enabled */
|
||||
__u64 enable_addr;
|
||||
|
||||
/* Input: Pointer to string with event name, description and flags */
|
||||
__u64 name_args;
|
||||
|
||||
/* Output: Index of the event to use when writing data */
|
||||
__u32 write_index;
|
||||
} __attribute__((__packed__));
|
||||
|
||||
/*
|
||||
* Describes an event unregister, callers must set the size, address and bit.
|
||||
* This structure is passed to the DIAG_IOCSUNREG ioctl to disable bit updates.
|
||||
*/
|
||||
struct user_unreg {
|
||||
/* Input: Size of the user_unreg structure being used */
|
||||
__u32 size;
|
||||
|
||||
/* Input: Bit to unregister */
|
||||
__u8 disable_bit;
|
||||
|
||||
/* Input: Reserved, set to 0 */
|
||||
__u8 __reserved;
|
||||
|
||||
/* Input: Reserved, set to 0 */
|
||||
__u16 __reserved2;
|
||||
|
||||
/* Input: Address to unregister */
|
||||
__u64 disable_addr;
|
||||
} __attribute__((__packed__));
|
||||
|
||||
#define DIAG_IOC_MAGIC '*'
|
||||
|
||||
/* Request to register a user_event */
|
||||
#define DIAG_IOCSREG _IOWR(DIAG_IOC_MAGIC, 0, struct user_reg *)
|
||||
|
||||
/* Request to delete a user_event */
|
||||
#define DIAG_IOCSDEL _IOW(DIAG_IOC_MAGIC, 1, char *)
|
||||
|
||||
/* Requests to unregister a user_event */
|
||||
#define DIAG_IOCSUNREG _IOW(DIAG_IOC_MAGIC, 2, struct user_unreg*)
|
||||
|
||||
#endif /* _UAPI_LINUX_USER_EVENTS_H */
|
@ -68,6 +68,7 @@
|
||||
#include <linux/kprobes.h>
|
||||
#include <linux/rethook.h>
|
||||
#include <linux/sysfs.h>
|
||||
#include <linux/user_events.h>
|
||||
|
||||
#include <linux/uaccess.h>
|
||||
#include <asm/unistd.h>
|
||||
@ -818,6 +819,7 @@ void __noreturn do_exit(long code)
|
||||
|
||||
coredump_task_exit(tsk);
|
||||
ptrace_event(PTRACE_EVENT_EXIT, code);
|
||||
user_events_exit(tsk);
|
||||
|
||||
validate_creds_for_do_exit(tsk);
|
||||
|
||||
|
@ -97,6 +97,7 @@
|
||||
#include <linux/io_uring.h>
|
||||
#include <linux/bpf.h>
|
||||
#include <linux/stackprotector.h>
|
||||
#include <linux/user_events.h>
|
||||
|
||||
#include <asm/pgalloc.h>
|
||||
#include <linux/uaccess.h>
|
||||
@ -2735,6 +2736,7 @@ __latent_entropy struct task_struct *copy_process(
|
||||
|
||||
trace_task_newtask(p, clone_flags);
|
||||
uprobe_copy_process(p, clone_flags);
|
||||
user_events_fork(p, clone_flags);
|
||||
|
||||
copy_oom_score_adj(clone_flags, p);
|
||||
|
||||
|
@ -792,15 +792,15 @@ config USER_EVENTS
|
||||
bool "User trace events"
|
||||
select TRACING
|
||||
select DYNAMIC_EVENTS
|
||||
depends on BROKEN || COMPILE_TEST # API needs to be straighten out
|
||||
help
|
||||
User trace events are user-defined trace events that
|
||||
can be used like an existing kernel trace event. User trace
|
||||
events are generated by writing to a tracefs file. User
|
||||
processes can determine if their tracing events should be
|
||||
generated by memory mapping a tracefs file and checking for
|
||||
an associated byte being non-zero.
|
||||
generated by registering a value and bit with the kernel
|
||||
that reflects when it is enabled or not.
|
||||
|
||||
See Documentation/trace/user_events.rst.
|
||||
If in doubt, say N.
|
||||
|
||||
config HIST_TRIGGERS
|
||||
|
@ -2640,9 +2640,20 @@ kprobe_multi_link_prog_run(struct bpf_kprobe_multi_link *link,
|
||||
return err;
|
||||
}
|
||||
|
||||
static void
|
||||
static int
|
||||
kprobe_multi_link_handler(struct fprobe *fp, unsigned long fentry_ip,
|
||||
struct pt_regs *regs)
|
||||
struct pt_regs *regs, void *data)
|
||||
{
|
||||
struct bpf_kprobe_multi_link *link;
|
||||
|
||||
link = container_of(fp, struct bpf_kprobe_multi_link, fp);
|
||||
kprobe_multi_link_prog_run(link, get_entry_ip(fentry_ip), regs);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void
|
||||
kprobe_multi_link_exit_handler(struct fprobe *fp, unsigned long fentry_ip,
|
||||
struct pt_regs *regs, void *data)
|
||||
{
|
||||
struct bpf_kprobe_multi_link *link;
|
||||
|
||||
@ -2844,7 +2855,7 @@ int bpf_kprobe_multi_link_attach(const union bpf_attr *attr, struct bpf_prog *pr
|
||||
goto error;
|
||||
|
||||
if (flags & BPF_F_KPROBE_MULTI_RETURN)
|
||||
link->fp.exit_handler = kprobe_multi_link_handler;
|
||||
link->fp.exit_handler = kprobe_multi_link_exit_handler;
|
||||
else
|
||||
link->fp.entry_handler = kprobe_multi_link_handler;
|
||||
|
||||
|
@ -17,15 +17,17 @@
|
||||
struct fprobe_rethook_node {
|
||||
struct rethook_node node;
|
||||
unsigned long entry_ip;
|
||||
char data[];
|
||||
};
|
||||
|
||||
static void fprobe_handler(unsigned long ip, unsigned long parent_ip,
|
||||
struct ftrace_ops *ops, struct ftrace_regs *fregs)
|
||||
{
|
||||
struct fprobe_rethook_node *fpr;
|
||||
struct rethook_node *rh;
|
||||
struct rethook_node *rh = NULL;
|
||||
struct fprobe *fp;
|
||||
int bit;
|
||||
void *entry_data = NULL;
|
||||
int bit, ret;
|
||||
|
||||
fp = container_of(ops, struct fprobe, ops);
|
||||
if (fprobe_disabled(fp))
|
||||
@ -37,9 +39,6 @@ static void fprobe_handler(unsigned long ip, unsigned long parent_ip,
|
||||
return;
|
||||
}
|
||||
|
||||
if (fp->entry_handler)
|
||||
fp->entry_handler(fp, ip, ftrace_get_regs(fregs));
|
||||
|
||||
if (fp->exit_handler) {
|
||||
rh = rethook_try_get(fp->rethook);
|
||||
if (!rh) {
|
||||
@ -48,9 +47,20 @@ static void fprobe_handler(unsigned long ip, unsigned long parent_ip,
|
||||
}
|
||||
fpr = container_of(rh, struct fprobe_rethook_node, node);
|
||||
fpr->entry_ip = ip;
|
||||
rethook_hook(rh, ftrace_get_regs(fregs), true);
|
||||
if (fp->entry_data_size)
|
||||
entry_data = fpr->data;
|
||||
}
|
||||
|
||||
if (fp->entry_handler)
|
||||
ret = fp->entry_handler(fp, ip, ftrace_get_regs(fregs), entry_data);
|
||||
|
||||
/* If entry_handler returns !0, nmissed is not counted. */
|
||||
if (rh) {
|
||||
if (ret)
|
||||
rethook_recycle(rh);
|
||||
else
|
||||
rethook_hook(rh, ftrace_get_regs(fregs), true);
|
||||
}
|
||||
out:
|
||||
ftrace_test_recursion_unlock(bit);
|
||||
}
|
||||
@ -81,7 +91,8 @@ static void fprobe_exit_handler(struct rethook_node *rh, void *data,
|
||||
|
||||
fpr = container_of(rh, struct fprobe_rethook_node, node);
|
||||
|
||||
fp->exit_handler(fp, fpr->entry_ip, regs);
|
||||
fp->exit_handler(fp, fpr->entry_ip, regs,
|
||||
fp->entry_data_size ? (void *)fpr->data : NULL);
|
||||
}
|
||||
NOKPROBE_SYMBOL(fprobe_exit_handler);
|
||||
|
||||
@ -136,6 +147,9 @@ static int fprobe_init_rethook(struct fprobe *fp, int num)
|
||||
}
|
||||
|
||||
/* Initialize rethook if needed */
|
||||
if (fp->nr_maxactive)
|
||||
size = fp->nr_maxactive;
|
||||
else
|
||||
size = num * num_possible_cpus() * 2;
|
||||
if (size < 0)
|
||||
return -E2BIG;
|
||||
@ -146,7 +160,7 @@ static int fprobe_init_rethook(struct fprobe *fp, int num)
|
||||
for (i = 0; i < size; i++) {
|
||||
struct fprobe_rethook_node *node;
|
||||
|
||||
node = kzalloc(sizeof(*node), GFP_KERNEL);
|
||||
node = kzalloc(sizeof(*node) + fp->entry_data_size, GFP_KERNEL);
|
||||
if (!node) {
|
||||
rethook_free(fp->rethook);
|
||||
fp->rethook = NULL;
|
||||
|
@ -45,6 +45,9 @@
|
||||
#include "trace_output.h"
|
||||
#include "trace_stat.h"
|
||||
|
||||
/* Flags that do not get reset */
|
||||
#define FTRACE_NOCLEAR_FLAGS (FTRACE_FL_DISABLED | FTRACE_FL_TOUCHED)
|
||||
|
||||
#define FTRACE_INVALID_FUNCTION "__ftrace_invalid_address__"
|
||||
|
||||
#define FTRACE_WARN_ON(cond) \
|
||||
@ -2256,7 +2259,7 @@ static int ftrace_check_record(struct dyn_ftrace *rec, bool enable, bool update)
|
||||
flag ^= rec->flags & FTRACE_FL_ENABLED;
|
||||
|
||||
if (update) {
|
||||
rec->flags |= FTRACE_FL_ENABLED;
|
||||
rec->flags |= FTRACE_FL_ENABLED | FTRACE_FL_TOUCHED;
|
||||
if (flag & FTRACE_FL_REGS) {
|
||||
if (rec->flags & FTRACE_FL_REGS)
|
||||
rec->flags |= FTRACE_FL_REGS_EN;
|
||||
@ -2326,7 +2329,7 @@ static int ftrace_check_record(struct dyn_ftrace *rec, bool enable, bool update)
|
||||
if (update) {
|
||||
/* If there's no more users, clear all flags */
|
||||
if (!ftrace_rec_count(rec))
|
||||
rec->flags &= FTRACE_FL_DISABLED;
|
||||
rec->flags &= FTRACE_NOCLEAR_FLAGS;
|
||||
else
|
||||
/*
|
||||
* Just disable the record, but keep the ops TRAMP
|
||||
@ -3147,7 +3150,7 @@ int ftrace_shutdown(struct ftrace_ops *ops, int command)
|
||||
struct dyn_ftrace *rec;
|
||||
|
||||
do_for_each_ftrace_rec(pg, rec) {
|
||||
if (FTRACE_WARN_ON_ONCE(rec->flags & ~FTRACE_FL_DISABLED))
|
||||
if (FTRACE_WARN_ON_ONCE(rec->flags & ~FTRACE_NOCLEAR_FLAGS))
|
||||
pr_warn(" %pS flags:%lx\n",
|
||||
(void *)rec->ip, rec->flags);
|
||||
} while_for_each_ftrace_rec();
|
||||
@ -3598,7 +3601,10 @@ t_func_next(struct seq_file *m, loff_t *pos)
|
||||
!ftrace_lookup_ip(iter->hash, rec->ip)) ||
|
||||
|
||||
((iter->flags & FTRACE_ITER_ENABLED) &&
|
||||
!(rec->flags & FTRACE_FL_ENABLED))) {
|
||||
!(rec->flags & FTRACE_FL_ENABLED)) ||
|
||||
|
||||
((iter->flags & FTRACE_ITER_TOUCHED) &&
|
||||
!(rec->flags & FTRACE_FL_TOUCHED))) {
|
||||
|
||||
rec = NULL;
|
||||
goto retry;
|
||||
@ -3857,7 +3863,7 @@ static int t_show(struct seq_file *m, void *v)
|
||||
return 0;
|
||||
}
|
||||
|
||||
if (iter->flags & FTRACE_ITER_ENABLED) {
|
||||
if (iter->flags & (FTRACE_ITER_ENABLED | FTRACE_ITER_TOUCHED)) {
|
||||
struct ftrace_ops *ops;
|
||||
|
||||
seq_printf(m, " (%ld)%s%s%s%s",
|
||||
@ -3959,6 +3965,31 @@ ftrace_enabled_open(struct inode *inode, struct file *file)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
ftrace_touched_open(struct inode *inode, struct file *file)
|
||||
{
|
||||
struct ftrace_iterator *iter;
|
||||
|
||||
/*
|
||||
* This shows us what functions have ever been enabled
|
||||
* (traced, direct, patched, etc). Not sure if we want lockdown
|
||||
* to hide such critical information for an admin.
|
||||
* Although, perhaps it can show information we don't
|
||||
* want people to see, but if something had traced
|
||||
* something, we probably want to know about it.
|
||||
*/
|
||||
|
||||
iter = __seq_open_private(file, &show_ftrace_seq_ops, sizeof(*iter));
|
||||
if (!iter)
|
||||
return -ENOMEM;
|
||||
|
||||
iter->pg = ftrace_pages_start;
|
||||
iter->flags = FTRACE_ITER_TOUCHED;
|
||||
iter->ops = &global_ops;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* ftrace_regex_open - initialize function tracer filter files
|
||||
* @ops: The ftrace_ops that hold the hash filters
|
||||
@ -5872,6 +5903,13 @@ static const struct file_operations ftrace_enabled_fops = {
|
||||
.release = seq_release_private,
|
||||
};
|
||||
|
||||
static const struct file_operations ftrace_touched_fops = {
|
||||
.open = ftrace_touched_open,
|
||||
.read = seq_read,
|
||||
.llseek = seq_lseek,
|
||||
.release = seq_release_private,
|
||||
};
|
||||
|
||||
static const struct file_operations ftrace_filter_fops = {
|
||||
.open = ftrace_filter_open,
|
||||
.read = seq_read,
|
||||
@ -6336,6 +6374,9 @@ static __init int ftrace_init_dyn_tracefs(struct dentry *d_tracer)
|
||||
trace_create_file("enabled_functions", TRACE_MODE_READ,
|
||||
d_tracer, NULL, &ftrace_enabled_fops);
|
||||
|
||||
trace_create_file("touched_functions", TRACE_MODE_READ,
|
||||
d_tracer, NULL, &ftrace_touched_fops);
|
||||
|
||||
ftrace_create_filter_files(&global_ops, d_tracer);
|
||||
|
||||
#ifdef CONFIG_FUNCTION_GRAPH_TRACER
|
||||
|
@ -163,7 +163,7 @@ enum {
|
||||
#define extended_time(event) \
|
||||
(event->type_len >= RINGBUF_TYPE_TIME_EXTEND)
|
||||
|
||||
static inline int rb_null_event(struct ring_buffer_event *event)
|
||||
static inline bool rb_null_event(struct ring_buffer_event *event)
|
||||
{
|
||||
return event->type_len == RINGBUF_TYPE_PADDING && !event->time_delta;
|
||||
}
|
||||
@ -363,11 +363,9 @@ static void free_buffer_page(struct buffer_page *bpage)
|
||||
/*
|
||||
* We need to fit the time_stamp delta into 27 bits.
|
||||
*/
|
||||
static inline int test_time_stamp(u64 delta)
|
||||
static inline bool test_time_stamp(u64 delta)
|
||||
{
|
||||
if (delta & TS_DELTA_TEST)
|
||||
return 1;
|
||||
return 0;
|
||||
return !!(delta & TS_DELTA_TEST);
|
||||
}
|
||||
|
||||
#define BUF_PAGE_SIZE (PAGE_SIZE - BUF_PAGE_HDR_SIZE)
|
||||
@ -696,7 +694,7 @@ rb_time_read_cmpxchg(local_t *l, unsigned long expect, unsigned long set)
|
||||
return ret == expect;
|
||||
}
|
||||
|
||||
static int rb_time_cmpxchg(rb_time_t *t, u64 expect, u64 set)
|
||||
static bool rb_time_cmpxchg(rb_time_t *t, u64 expect, u64 set)
|
||||
{
|
||||
unsigned long cnt, top, bottom, msb;
|
||||
unsigned long cnt2, top2, bottom2, msb2;
|
||||
@ -1486,7 +1484,7 @@ rb_set_head_page(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static int rb_head_page_replace(struct buffer_page *old,
|
||||
static bool rb_head_page_replace(struct buffer_page *old,
|
||||
struct buffer_page *new)
|
||||
{
|
||||
unsigned long *ptr = (unsigned long *)&old->list.prev->next;
|
||||
@ -1565,15 +1563,12 @@ static void rb_tail_page_update(struct ring_buffer_per_cpu *cpu_buffer,
|
||||
}
|
||||
}
|
||||
|
||||
static int rb_check_bpage(struct ring_buffer_per_cpu *cpu_buffer,
|
||||
static void rb_check_bpage(struct ring_buffer_per_cpu *cpu_buffer,
|
||||
struct buffer_page *bpage)
|
||||
{
|
||||
unsigned long val = (unsigned long)bpage;
|
||||
|
||||
if (RB_WARN_ON(cpu_buffer, val & RB_FLAG_MASK))
|
||||
return 1;
|
||||
|
||||
return 0;
|
||||
RB_WARN_ON(cpu_buffer, val & RB_FLAG_MASK);
|
||||
}
|
||||
|
||||
/**
|
||||
@ -1583,30 +1578,28 @@ static int rb_check_bpage(struct ring_buffer_per_cpu *cpu_buffer,
|
||||
* As a safety measure we check to make sure the data pages have not
|
||||
* been corrupted.
|
||||
*/
|
||||
static int rb_check_pages(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
static void rb_check_pages(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
{
|
||||
struct list_head *head = rb_list_head(cpu_buffer->pages);
|
||||
struct list_head *tmp;
|
||||
|
||||
if (RB_WARN_ON(cpu_buffer,
|
||||
rb_list_head(rb_list_head(head->next)->prev) != head))
|
||||
return -1;
|
||||
return;
|
||||
|
||||
if (RB_WARN_ON(cpu_buffer,
|
||||
rb_list_head(rb_list_head(head->prev)->next) != head))
|
||||
return -1;
|
||||
return;
|
||||
|
||||
for (tmp = rb_list_head(head->next); tmp != head; tmp = rb_list_head(tmp->next)) {
|
||||
if (RB_WARN_ON(cpu_buffer,
|
||||
rb_list_head(rb_list_head(tmp->next)->prev) != tmp))
|
||||
return -1;
|
||||
return;
|
||||
|
||||
if (RB_WARN_ON(cpu_buffer,
|
||||
rb_list_head(rb_list_head(tmp->prev)->next) != tmp))
|
||||
return -1;
|
||||
return;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int __rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer,
|
||||
@ -1774,6 +1767,8 @@ static void rb_free_cpu_buffer(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
struct list_head *head = cpu_buffer->pages;
|
||||
struct buffer_page *bpage, *tmp;
|
||||
|
||||
irq_work_sync(&cpu_buffer->irq_work.work);
|
||||
|
||||
free_buffer_page(cpu_buffer->reader_page);
|
||||
|
||||
if (head) {
|
||||
@ -1880,6 +1875,8 @@ ring_buffer_free(struct trace_buffer *buffer)
|
||||
|
||||
cpuhp_state_remove_instance(CPUHP_TRACE_RB_PREPARE, &buffer->node);
|
||||
|
||||
irq_work_sync(&buffer->irq_work.work);
|
||||
|
||||
for_each_buffer_cpu(buffer, cpu)
|
||||
rb_free_cpu_buffer(buffer->buffers[cpu]);
|
||||
|
||||
@ -1918,7 +1915,7 @@ static inline unsigned long rb_page_write(struct buffer_page *bpage)
|
||||
return local_read(&bpage->write) & RB_WRITE_MASK;
|
||||
}
|
||||
|
||||
static int
|
||||
static bool
|
||||
rb_remove_pages(struct ring_buffer_per_cpu *cpu_buffer, unsigned long nr_pages)
|
||||
{
|
||||
struct list_head *tail_page, *to_remove, *next_page;
|
||||
@ -2031,12 +2028,13 @@ rb_remove_pages(struct ring_buffer_per_cpu *cpu_buffer, unsigned long nr_pages)
|
||||
return nr_removed == 0;
|
||||
}
|
||||
|
||||
static int
|
||||
static bool
|
||||
rb_insert_pages(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
{
|
||||
struct list_head *pages = &cpu_buffer->new_pages;
|
||||
int retries, success;
|
||||
unsigned long flags;
|
||||
bool success;
|
||||
int retries;
|
||||
|
||||
/* Can be called at early boot up, where interrupts must not been enabled */
|
||||
raw_spin_lock_irqsave(&cpu_buffer->reader_lock, flags);
|
||||
@ -2055,15 +2053,16 @@ rb_insert_pages(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
* spinning.
|
||||
*/
|
||||
retries = 10;
|
||||
success = 0;
|
||||
success = false;
|
||||
while (retries--) {
|
||||
struct list_head *head_page, *prev_page, *r;
|
||||
struct list_head *last_page, *first_page;
|
||||
struct list_head *head_page_with_bit;
|
||||
struct buffer_page *hpage = rb_set_head_page(cpu_buffer);
|
||||
|
||||
head_page = &rb_set_head_page(cpu_buffer)->list;
|
||||
if (!head_page)
|
||||
if (!hpage)
|
||||
break;
|
||||
head_page = &hpage->list;
|
||||
prev_page = head_page->prev;
|
||||
|
||||
first_page = pages->next;
|
||||
@ -2084,7 +2083,7 @@ rb_insert_pages(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
* pointer to point to end of list
|
||||
*/
|
||||
head_page->prev = last_page;
|
||||
success = 1;
|
||||
success = true;
|
||||
break;
|
||||
}
|
||||
}
|
||||
@ -2112,7 +2111,7 @@ rb_insert_pages(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
|
||||
static void rb_update_pages(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
{
|
||||
int success;
|
||||
bool success;
|
||||
|
||||
if (cpu_buffer->nr_pages_to_update > 0)
|
||||
success = rb_insert_pages(cpu_buffer);
|
||||
@ -2995,7 +2994,7 @@ static u64 rb_time_delta(struct ring_buffer_event *event)
|
||||
}
|
||||
}
|
||||
|
||||
static inline int
|
||||
static inline bool
|
||||
rb_try_to_discard(struct ring_buffer_per_cpu *cpu_buffer,
|
||||
struct ring_buffer_event *event)
|
||||
{
|
||||
@ -3016,7 +3015,7 @@ rb_try_to_discard(struct ring_buffer_per_cpu *cpu_buffer,
|
||||
delta = rb_time_delta(event);
|
||||
|
||||
if (!rb_time_read(&cpu_buffer->write_stamp, &write_stamp))
|
||||
return 0;
|
||||
return false;
|
||||
|
||||
/* Make sure the write stamp is read before testing the location */
|
||||
barrier();
|
||||
@ -3029,7 +3028,7 @@ rb_try_to_discard(struct ring_buffer_per_cpu *cpu_buffer,
|
||||
/* Something came in, can't discard */
|
||||
if (!rb_time_cmpxchg(&cpu_buffer->write_stamp,
|
||||
write_stamp, write_stamp - delta))
|
||||
return 0;
|
||||
return false;
|
||||
|
||||
/*
|
||||
* It's possible that the event time delta is zero
|
||||
@ -3062,12 +3061,12 @@ rb_try_to_discard(struct ring_buffer_per_cpu *cpu_buffer,
|
||||
if (index == old_index) {
|
||||
/* update counters */
|
||||
local_sub(event_length, &cpu_buffer->entries_bytes);
|
||||
return 1;
|
||||
return true;
|
||||
}
|
||||
}
|
||||
|
||||
/* could not discard */
|
||||
return 0;
|
||||
return false;
|
||||
}
|
||||
|
||||
static void rb_start_commit(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
@ -3288,7 +3287,7 @@ rb_wakeups(struct trace_buffer *buffer, struct ring_buffer_per_cpu *cpu_buffer)
|
||||
* Note: The TRANSITION bit only handles a single transition between context.
|
||||
*/
|
||||
|
||||
static __always_inline int
|
||||
static __always_inline bool
|
||||
trace_recursive_lock(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
{
|
||||
unsigned int val = cpu_buffer->current_context;
|
||||
@ -3305,14 +3304,14 @@ trace_recursive_lock(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
bit = RB_CTX_TRANSITION;
|
||||
if (val & (1 << (bit + cpu_buffer->nest))) {
|
||||
do_ring_buffer_record_recursion();
|
||||
return 1;
|
||||
return true;
|
||||
}
|
||||
}
|
||||
|
||||
val |= (1 << (bit + cpu_buffer->nest));
|
||||
cpu_buffer->current_context = val;
|
||||
|
||||
return 0;
|
||||
return false;
|
||||
}
|
||||
|
||||
static __always_inline void
|
||||
@ -4069,10 +4068,10 @@ void ring_buffer_record_off(struct trace_buffer *buffer)
|
||||
unsigned int rd;
|
||||
unsigned int new_rd;
|
||||
|
||||
do {
|
||||
rd = atomic_read(&buffer->record_disabled);
|
||||
do {
|
||||
new_rd = rd | RB_BUFFER_OFF;
|
||||
} while (atomic_cmpxchg(&buffer->record_disabled, rd, new_rd) != rd);
|
||||
} while (!atomic_try_cmpxchg(&buffer->record_disabled, &rd, new_rd));
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(ring_buffer_record_off);
|
||||
|
||||
@ -4092,10 +4091,10 @@ void ring_buffer_record_on(struct trace_buffer *buffer)
|
||||
unsigned int rd;
|
||||
unsigned int new_rd;
|
||||
|
||||
do {
|
||||
rd = atomic_read(&buffer->record_disabled);
|
||||
do {
|
||||
new_rd = rd & ~RB_BUFFER_OFF;
|
||||
} while (atomic_cmpxchg(&buffer->record_disabled, rd, new_rd) != rd);
|
||||
} while (!atomic_try_cmpxchg(&buffer->record_disabled, &rd, new_rd));
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(ring_buffer_record_on);
|
||||
|
||||
@ -4502,7 +4501,6 @@ rb_update_read_stamp(struct ring_buffer_per_cpu *cpu_buffer,
|
||||
default:
|
||||
RB_WARN_ON(cpu_buffer, 1);
|
||||
}
|
||||
return;
|
||||
}
|
||||
|
||||
static void
|
||||
@ -4533,7 +4531,6 @@ rb_update_iter_read_stamp(struct ring_buffer_iter *iter,
|
||||
default:
|
||||
RB_WARN_ON(iter->cpu_buffer, 1);
|
||||
}
|
||||
return;
|
||||
}
|
||||
|
||||
static struct buffer_page *
|
||||
@ -4543,7 +4540,7 @@ rb_get_reader_page(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
unsigned long overwrite;
|
||||
unsigned long flags;
|
||||
int nr_loops = 0;
|
||||
int ret;
|
||||
bool ret;
|
||||
|
||||
local_irq_save(flags);
|
||||
arch_spin_lock(&cpu_buffer->lock);
|
||||
@ -4953,7 +4950,6 @@ rb_reader_unlock(struct ring_buffer_per_cpu *cpu_buffer, bool locked)
|
||||
{
|
||||
if (likely(locked))
|
||||
raw_spin_unlock(&cpu_buffer->reader_lock);
|
||||
return;
|
||||
}
|
||||
|
||||
/**
|
||||
@ -5345,6 +5341,9 @@ void ring_buffer_reset_cpu(struct trace_buffer *buffer, int cpu)
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(ring_buffer_reset_cpu);
|
||||
|
||||
/* Flag to ensure proper resetting of atomic variables */
|
||||
#define RESET_BIT (1 << 30)
|
||||
|
||||
/**
|
||||
* ring_buffer_reset_online_cpus - reset a ring buffer per CPU buffer
|
||||
* @buffer: The ring buffer to reset a per cpu buffer of
|
||||
@ -5361,20 +5360,27 @@ void ring_buffer_reset_online_cpus(struct trace_buffer *buffer)
|
||||
for_each_online_buffer_cpu(buffer, cpu) {
|
||||
cpu_buffer = buffer->buffers[cpu];
|
||||
|
||||
atomic_inc(&cpu_buffer->resize_disabled);
|
||||
atomic_add(RESET_BIT, &cpu_buffer->resize_disabled);
|
||||
atomic_inc(&cpu_buffer->record_disabled);
|
||||
}
|
||||
|
||||
/* Make sure all commits have finished */
|
||||
synchronize_rcu();
|
||||
|
||||
for_each_online_buffer_cpu(buffer, cpu) {
|
||||
for_each_buffer_cpu(buffer, cpu) {
|
||||
cpu_buffer = buffer->buffers[cpu];
|
||||
|
||||
/*
|
||||
* If a CPU came online during the synchronize_rcu(), then
|
||||
* ignore it.
|
||||
*/
|
||||
if (!(atomic_read(&cpu_buffer->resize_disabled) & RESET_BIT))
|
||||
continue;
|
||||
|
||||
reset_disabled_cpu_buffer(cpu_buffer);
|
||||
|
||||
atomic_dec(&cpu_buffer->record_disabled);
|
||||
atomic_dec(&cpu_buffer->resize_disabled);
|
||||
atomic_sub(RESET_BIT, &cpu_buffer->resize_disabled);
|
||||
}
|
||||
|
||||
mutex_unlock(&buffer->mutex);
|
||||
@ -5424,8 +5430,8 @@ bool ring_buffer_empty(struct trace_buffer *buffer)
|
||||
struct ring_buffer_per_cpu *cpu_buffer;
|
||||
unsigned long flags;
|
||||
bool dolock;
|
||||
bool ret;
|
||||
int cpu;
|
||||
int ret;
|
||||
|
||||
/* yes this is racy, but if you don't like the race, lock the buffer */
|
||||
for_each_buffer_cpu(buffer, cpu) {
|
||||
@ -5454,7 +5460,7 @@ bool ring_buffer_empty_cpu(struct trace_buffer *buffer, int cpu)
|
||||
struct ring_buffer_per_cpu *cpu_buffer;
|
||||
unsigned long flags;
|
||||
bool dolock;
|
||||
int ret;
|
||||
bool ret;
|
||||
|
||||
if (!cpumask_test_cpu(cpu, buffer->cpumask))
|
||||
return true;
|
||||
|
@ -3726,7 +3726,7 @@ __find_next_entry(struct trace_iterator *iter, int *ent_cpu,
|
||||
#define STATIC_FMT_BUF_SIZE 128
|
||||
static char static_fmt_buf[STATIC_FMT_BUF_SIZE];
|
||||
|
||||
static char *trace_iter_expand_format(struct trace_iterator *iter)
|
||||
char *trace_iter_expand_format(struct trace_iterator *iter)
|
||||
{
|
||||
char *tmp;
|
||||
|
||||
@ -4446,8 +4446,11 @@ static enum print_line_t print_trace_fmt(struct trace_iterator *iter)
|
||||
if (trace_seq_has_overflowed(s))
|
||||
return TRACE_TYPE_PARTIAL_LINE;
|
||||
|
||||
if (event)
|
||||
if (event) {
|
||||
if (tr->trace_flags & TRACE_ITER_FIELDS)
|
||||
return print_event_fields(iter, event);
|
||||
return event->funcs->trace(iter, sym_flags, event);
|
||||
}
|
||||
|
||||
trace_seq_printf(s, "Unknown type %d\n", entry->type);
|
||||
|
||||
|
@ -619,6 +619,7 @@ bool trace_is_tracepoint_string(const char *str);
|
||||
const char *trace_event_format(struct trace_iterator *iter, const char *fmt);
|
||||
void trace_check_vprintf(struct trace_iterator *iter, const char *fmt,
|
||||
va_list ap) __printf(2, 0);
|
||||
char *trace_iter_expand_format(struct trace_iterator *iter);
|
||||
|
||||
int trace_empty(struct trace_iterator *iter);
|
||||
|
||||
@ -1199,6 +1200,7 @@ extern int trace_get_user(struct trace_parser *parser, const char __user *ubuf,
|
||||
C(HEX, "hex"), \
|
||||
C(BIN, "bin"), \
|
||||
C(BLOCK, "block"), \
|
||||
C(FIELDS, "fields"), \
|
||||
C(PRINTK, "trace_printk"), \
|
||||
C(ANNOTATE, "annotate"), \
|
||||
C(USERSTACKTRACE, "userstacktrace"), \
|
||||
|
File diff suppressed because it is too large
Load Diff
@ -221,8 +221,11 @@ trace_print_hex_seq(struct trace_seq *p, const unsigned char *buf, int buf_len,
|
||||
const char *ret = trace_seq_buffer_ptr(p);
|
||||
const char *fmt = concatenate ? "%*phN" : "%*ph";
|
||||
|
||||
for (i = 0; i < buf_len; i += 16)
|
||||
for (i = 0; i < buf_len; i += 16) {
|
||||
if (!concatenate && i != 0)
|
||||
trace_seq_putc(p, ' ');
|
||||
trace_seq_printf(p, fmt, min(buf_len - i, 16), &buf[i]);
|
||||
}
|
||||
trace_seq_putc(p, 0);
|
||||
|
||||
return ret;
|
||||
@ -808,6 +811,176 @@ EXPORT_SYMBOL_GPL(unregister_trace_event);
|
||||
* Standard events
|
||||
*/
|
||||
|
||||
static void print_array(struct trace_iterator *iter, void *pos,
|
||||
struct ftrace_event_field *field)
|
||||
{
|
||||
int offset;
|
||||
int len;
|
||||
int i;
|
||||
|
||||
offset = *(int *)pos & 0xffff;
|
||||
len = *(int *)pos >> 16;
|
||||
|
||||
if (field)
|
||||
offset += field->offset + sizeof(int);
|
||||
|
||||
if (offset + len > iter->ent_size) {
|
||||
trace_seq_puts(&iter->seq, "<OVERFLOW>");
|
||||
return;
|
||||
}
|
||||
|
||||
pos = (void *)iter->ent + offset;
|
||||
|
||||
for (i = 0; i < len; i++, pos++) {
|
||||
if (i)
|
||||
trace_seq_putc(&iter->seq, ',');
|
||||
trace_seq_printf(&iter->seq, "%02x", *(unsigned char *)pos);
|
||||
}
|
||||
}
|
||||
|
||||
static void print_fields(struct trace_iterator *iter, struct trace_event_call *call,
|
||||
struct list_head *head)
|
||||
{
|
||||
struct ftrace_event_field *field;
|
||||
int offset;
|
||||
int len;
|
||||
int ret;
|
||||
void *pos;
|
||||
|
||||
list_for_each_entry(field, head, link) {
|
||||
trace_seq_printf(&iter->seq, " %s=", field->name);
|
||||
if (field->offset + field->size > iter->ent_size) {
|
||||
trace_seq_puts(&iter->seq, "<OVERFLOW>");
|
||||
continue;
|
||||
}
|
||||
pos = (void *)iter->ent + field->offset;
|
||||
|
||||
switch (field->filter_type) {
|
||||
case FILTER_COMM:
|
||||
case FILTER_STATIC_STRING:
|
||||
trace_seq_printf(&iter->seq, "%.*s", field->size, (char *)pos);
|
||||
break;
|
||||
case FILTER_RDYN_STRING:
|
||||
case FILTER_DYN_STRING:
|
||||
offset = *(int *)pos & 0xffff;
|
||||
len = *(int *)pos >> 16;
|
||||
|
||||
if (field->filter_type == FILTER_RDYN_STRING)
|
||||
offset += field->offset + sizeof(int);
|
||||
|
||||
if (offset + len > iter->ent_size) {
|
||||
trace_seq_puts(&iter->seq, "<OVERFLOW>");
|
||||
break;
|
||||
}
|
||||
pos = (void *)iter->ent + offset;
|
||||
trace_seq_printf(&iter->seq, "%.*s", len, (char *)pos);
|
||||
break;
|
||||
case FILTER_PTR_STRING:
|
||||
if (!iter->fmt_size)
|
||||
trace_iter_expand_format(iter);
|
||||
pos = *(void **)pos;
|
||||
ret = strncpy_from_kernel_nofault(iter->fmt, pos,
|
||||
iter->fmt_size);
|
||||
if (ret < 0)
|
||||
trace_seq_printf(&iter->seq, "(0x%px)", pos);
|
||||
else
|
||||
trace_seq_printf(&iter->seq, "(0x%px:%s)",
|
||||
pos, iter->fmt);
|
||||
break;
|
||||
case FILTER_TRACE_FN:
|
||||
pos = *(void **)pos;
|
||||
trace_seq_printf(&iter->seq, "%pS", pos);
|
||||
break;
|
||||
case FILTER_CPU:
|
||||
case FILTER_OTHER:
|
||||
switch (field->size) {
|
||||
case 1:
|
||||
if (isprint(*(char *)pos)) {
|
||||
trace_seq_printf(&iter->seq, "'%c'",
|
||||
*(unsigned char *)pos);
|
||||
}
|
||||
trace_seq_printf(&iter->seq, "(%d)",
|
||||
*(unsigned char *)pos);
|
||||
break;
|
||||
case 2:
|
||||
trace_seq_printf(&iter->seq, "0x%x (%d)",
|
||||
*(unsigned short *)pos,
|
||||
*(unsigned short *)pos);
|
||||
break;
|
||||
case 4:
|
||||
/* dynamic array info is 4 bytes */
|
||||
if (strstr(field->type, "__data_loc")) {
|
||||
print_array(iter, pos, NULL);
|
||||
break;
|
||||
}
|
||||
|
||||
if (strstr(field->type, "__rel_loc")) {
|
||||
print_array(iter, pos, field);
|
||||
break;
|
||||
}
|
||||
|
||||
trace_seq_printf(&iter->seq, "0x%x (%d)",
|
||||
*(unsigned int *)pos,
|
||||
*(unsigned int *)pos);
|
||||
break;
|
||||
case 8:
|
||||
trace_seq_printf(&iter->seq, "0x%llx (%lld)",
|
||||
*(unsigned long long *)pos,
|
||||
*(unsigned long long *)pos);
|
||||
break;
|
||||
default:
|
||||
trace_seq_puts(&iter->seq, "<INVALID-SIZE>");
|
||||
break;
|
||||
}
|
||||
break;
|
||||
default:
|
||||
trace_seq_puts(&iter->seq, "<INVALID-TYPE>");
|
||||
}
|
||||
}
|
||||
trace_seq_putc(&iter->seq, '\n');
|
||||
}
|
||||
|
||||
enum print_line_t print_event_fields(struct trace_iterator *iter,
|
||||
struct trace_event *event)
|
||||
{
|
||||
struct trace_event_call *call;
|
||||
struct list_head *head;
|
||||
|
||||
/* ftrace defined events have separate call structures */
|
||||
if (event->type <= __TRACE_LAST_TYPE) {
|
||||
bool found = false;
|
||||
|
||||
down_read(&trace_event_sem);
|
||||
list_for_each_entry(call, &ftrace_events, list) {
|
||||
if (call->event.type == event->type) {
|
||||
found = true;
|
||||
break;
|
||||
}
|
||||
/* No need to search all events */
|
||||
if (call->event.type > __TRACE_LAST_TYPE)
|
||||
break;
|
||||
}
|
||||
up_read(&trace_event_sem);
|
||||
if (!found) {
|
||||
trace_seq_printf(&iter->seq, "UNKNOWN TYPE %d\n", event->type);
|
||||
goto out;
|
||||
}
|
||||
} else {
|
||||
call = container_of(event, struct trace_event_call, event);
|
||||
}
|
||||
head = trace_get_fields(call);
|
||||
|
||||
trace_seq_printf(&iter->seq, "%s:", trace_event_name(call));
|
||||
|
||||
if (head && !list_empty(head))
|
||||
print_fields(iter, call, head);
|
||||
else
|
||||
trace_seq_puts(&iter->seq, "No fields found\n");
|
||||
|
||||
out:
|
||||
return trace_handle_return(&iter->seq);
|
||||
}
|
||||
|
||||
enum print_line_t trace_nop_print(struct trace_iterator *iter, int flags,
|
||||
struct trace_event *event)
|
||||
{
|
||||
|
@ -19,6 +19,8 @@ seq_print_ip_sym(struct trace_seq *s, unsigned long ip,
|
||||
extern void trace_seq_print_sym(struct trace_seq *s, unsigned long address, bool offset);
|
||||
extern int trace_print_context(struct trace_iterator *iter);
|
||||
extern int trace_print_lat_context(struct trace_iterator *iter);
|
||||
extern enum print_line_t print_event_fields(struct trace_iterator *iter,
|
||||
struct trace_event *event);
|
||||
|
||||
extern void trace_event_read_lock(void);
|
||||
extern void trace_event_read_unlock(void);
|
||||
|
@ -93,6 +93,38 @@ int seq_buf_printf(struct seq_buf *s, const char *fmt, ...)
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(seq_buf_printf);
|
||||
|
||||
/**
|
||||
* seq_buf_do_printk - printk seq_buf line by line
|
||||
* @s: seq_buf descriptor
|
||||
* @lvl: printk level
|
||||
*
|
||||
* printk()-s a multi-line sequential buffer line by line. The function
|
||||
* makes sure that the buffer in @s is nul terminated and safe to read
|
||||
* as a string.
|
||||
*/
|
||||
void seq_buf_do_printk(struct seq_buf *s, const char *lvl)
|
||||
{
|
||||
const char *start, *lf;
|
||||
|
||||
if (s->size == 0 || s->len == 0)
|
||||
return;
|
||||
|
||||
seq_buf_terminate(s);
|
||||
|
||||
start = s->buffer;
|
||||
while ((lf = strchr(start, '\n'))) {
|
||||
int len = lf - start + 1;
|
||||
|
||||
printk("%s%.*s", lvl, len, start);
|
||||
start = ++lf;
|
||||
}
|
||||
|
||||
/* No trailing LF */
|
||||
if (start < s->buffer + s->len)
|
||||
printk("%s%s\n", lvl, start);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(seq_buf_do_printk);
|
||||
|
||||
#ifdef CONFIG_BINARY_PRINTF
|
||||
/**
|
||||
* seq_buf_bprintf - Write the printf string from binary arguments
|
||||
|
@ -17,8 +17,11 @@ static u32 rand1, entry_val, exit_val;
|
||||
/* Use indirect calls to avoid inlining the target functions */
|
||||
static u32 (*target)(u32 value);
|
||||
static u32 (*target2)(u32 value);
|
||||
static u32 (*target_nest)(u32 value, u32 (*nest)(u32));
|
||||
static unsigned long target_ip;
|
||||
static unsigned long target2_ip;
|
||||
static unsigned long target_nest_ip;
|
||||
static int entry_return_value;
|
||||
|
||||
static noinline u32 fprobe_selftest_target(u32 value)
|
||||
{
|
||||
@ -30,16 +33,31 @@ static noinline u32 fprobe_selftest_target2(u32 value)
|
||||
return (value / div_factor) + 1;
|
||||
}
|
||||
|
||||
static notrace void fp_entry_handler(struct fprobe *fp, unsigned long ip, struct pt_regs *regs)
|
||||
static noinline u32 fprobe_selftest_nest_target(u32 value, u32 (*nest)(u32))
|
||||
{
|
||||
return nest(value + 2);
|
||||
}
|
||||
|
||||
static notrace int fp_entry_handler(struct fprobe *fp, unsigned long ip,
|
||||
struct pt_regs *regs, void *data)
|
||||
{
|
||||
KUNIT_EXPECT_FALSE(current_test, preemptible());
|
||||
/* This can be called on the fprobe_selftest_target and the fprobe_selftest_target2 */
|
||||
if (ip != target_ip)
|
||||
KUNIT_EXPECT_EQ(current_test, ip, target2_ip);
|
||||
entry_val = (rand1 / div_factor);
|
||||
if (fp->entry_data_size) {
|
||||
KUNIT_EXPECT_NOT_NULL(current_test, data);
|
||||
if (data)
|
||||
*(u32 *)data = entry_val;
|
||||
} else
|
||||
KUNIT_EXPECT_NULL(current_test, data);
|
||||
|
||||
return entry_return_value;
|
||||
}
|
||||
|
||||
static notrace void fp_exit_handler(struct fprobe *fp, unsigned long ip, struct pt_regs *regs)
|
||||
static notrace void fp_exit_handler(struct fprobe *fp, unsigned long ip,
|
||||
struct pt_regs *regs, void *data)
|
||||
{
|
||||
unsigned long ret = regs_return_value(regs);
|
||||
|
||||
@ -51,6 +69,26 @@ static notrace void fp_exit_handler(struct fprobe *fp, unsigned long ip, struct
|
||||
KUNIT_EXPECT_EQ(current_test, ret, (rand1 / div_factor));
|
||||
KUNIT_EXPECT_EQ(current_test, entry_val, (rand1 / div_factor));
|
||||
exit_val = entry_val + div_factor;
|
||||
if (fp->entry_data_size) {
|
||||
KUNIT_EXPECT_NOT_NULL(current_test, data);
|
||||
if (data)
|
||||
KUNIT_EXPECT_EQ(current_test, *(u32 *)data, entry_val);
|
||||
} else
|
||||
KUNIT_EXPECT_NULL(current_test, data);
|
||||
}
|
||||
|
||||
static notrace int nest_entry_handler(struct fprobe *fp, unsigned long ip,
|
||||
struct pt_regs *regs, void *data)
|
||||
{
|
||||
KUNIT_EXPECT_FALSE(current_test, preemptible());
|
||||
return 0;
|
||||
}
|
||||
|
||||
static notrace void nest_exit_handler(struct fprobe *fp, unsigned long ip,
|
||||
struct pt_regs *regs, void *data)
|
||||
{
|
||||
KUNIT_EXPECT_FALSE(current_test, preemptible());
|
||||
KUNIT_EXPECT_EQ(current_test, ip, target_nest_ip);
|
||||
}
|
||||
|
||||
/* Test entry only (no rethook) */
|
||||
@ -132,6 +170,64 @@ static void test_fprobe_syms(struct kunit *test)
|
||||
KUNIT_EXPECT_EQ(test, 0, unregister_fprobe(&fp));
|
||||
}
|
||||
|
||||
/* Test private entry_data */
|
||||
static void test_fprobe_data(struct kunit *test)
|
||||
{
|
||||
struct fprobe fp = {
|
||||
.entry_handler = fp_entry_handler,
|
||||
.exit_handler = fp_exit_handler,
|
||||
.entry_data_size = sizeof(u32),
|
||||
};
|
||||
|
||||
current_test = test;
|
||||
KUNIT_EXPECT_EQ(test, 0, register_fprobe(&fp, "fprobe_selftest_target", NULL));
|
||||
|
||||
target(rand1);
|
||||
|
||||
KUNIT_EXPECT_EQ(test, 0, unregister_fprobe(&fp));
|
||||
}
|
||||
|
||||
/* Test nr_maxactive */
|
||||
static void test_fprobe_nest(struct kunit *test)
|
||||
{
|
||||
static const char *syms[] = {"fprobe_selftest_target", "fprobe_selftest_nest_target"};
|
||||
struct fprobe fp = {
|
||||
.entry_handler = nest_entry_handler,
|
||||
.exit_handler = nest_exit_handler,
|
||||
.nr_maxactive = 1,
|
||||
};
|
||||
|
||||
current_test = test;
|
||||
KUNIT_EXPECT_EQ(test, 0, register_fprobe_syms(&fp, syms, 2));
|
||||
|
||||
target_nest(rand1, target);
|
||||
KUNIT_EXPECT_EQ(test, 1, fp.nmissed);
|
||||
|
||||
KUNIT_EXPECT_EQ(test, 0, unregister_fprobe(&fp));
|
||||
}
|
||||
|
||||
static void test_fprobe_skip(struct kunit *test)
|
||||
{
|
||||
struct fprobe fp = {
|
||||
.entry_handler = fp_entry_handler,
|
||||
.exit_handler = fp_exit_handler,
|
||||
};
|
||||
|
||||
current_test = test;
|
||||
KUNIT_EXPECT_EQ(test, 0, register_fprobe(&fp, "fprobe_selftest_target", NULL));
|
||||
|
||||
entry_return_value = 1;
|
||||
entry_val = 0;
|
||||
exit_val = 0;
|
||||
target(rand1);
|
||||
KUNIT_EXPECT_NE(test, 0, entry_val);
|
||||
KUNIT_EXPECT_EQ(test, 0, exit_val);
|
||||
KUNIT_EXPECT_EQ(test, 0, fp.nmissed);
|
||||
entry_return_value = 0;
|
||||
|
||||
KUNIT_EXPECT_EQ(test, 0, unregister_fprobe(&fp));
|
||||
}
|
||||
|
||||
static unsigned long get_ftrace_location(void *func)
|
||||
{
|
||||
unsigned long size, addr = (unsigned long)func;
|
||||
@ -147,8 +243,10 @@ static int fprobe_test_init(struct kunit *test)
|
||||
rand1 = get_random_u32_above(div_factor);
|
||||
target = fprobe_selftest_target;
|
||||
target2 = fprobe_selftest_target2;
|
||||
target_nest = fprobe_selftest_nest_target;
|
||||
target_ip = get_ftrace_location(target);
|
||||
target2_ip = get_ftrace_location(target2);
|
||||
target_nest_ip = get_ftrace_location(target_nest);
|
||||
|
||||
return 0;
|
||||
}
|
||||
@ -157,6 +255,9 @@ static struct kunit_case fprobe_testcases[] = {
|
||||
KUNIT_CASE(test_fprobe_entry),
|
||||
KUNIT_CASE(test_fprobe),
|
||||
KUNIT_CASE(test_fprobe_syms),
|
||||
KUNIT_CASE(test_fprobe_data),
|
||||
KUNIT_CASE(test_fprobe_nest),
|
||||
KUNIT_CASE(test_fprobe_skip),
|
||||
{}
|
||||
};
|
||||
|
||||
|
@ -48,7 +48,8 @@ static void show_backtrace(void)
|
||||
stack_trace_print(stacks, len, 24);
|
||||
}
|
||||
|
||||
static void sample_entry_handler(struct fprobe *fp, unsigned long ip, struct pt_regs *regs)
|
||||
static int sample_entry_handler(struct fprobe *fp, unsigned long ip,
|
||||
struct pt_regs *regs, void *data)
|
||||
{
|
||||
if (use_trace)
|
||||
/*
|
||||
@ -61,9 +62,11 @@ static void sample_entry_handler(struct fprobe *fp, unsigned long ip, struct pt_
|
||||
nhit++;
|
||||
if (stackdump)
|
||||
show_backtrace();
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void sample_exit_handler(struct fprobe *fp, unsigned long ip, struct pt_regs *regs)
|
||||
static void sample_exit_handler(struct fprobe *fp, unsigned long ip, struct pt_regs *regs,
|
||||
void *data)
|
||||
{
|
||||
unsigned long rip = instruction_pointer(regs);
|
||||
|
||||
|
@ -9,51 +9,28 @@
|
||||
#include <errno.h>
|
||||
#include <sys/ioctl.h>
|
||||
#include <sys/mman.h>
|
||||
#include <sys/uio.h>
|
||||
#include <fcntl.h>
|
||||
#include <stdio.h>
|
||||
#include <unistd.h>
|
||||
#include <asm/bitsperlong.h>
|
||||
#include <endian.h>
|
||||
#include <linux/user_events.h>
|
||||
|
||||
#if __BITS_PER_LONG == 64
|
||||
#define endian_swap(x) htole64(x)
|
||||
#else
|
||||
#define endian_swap(x) htole32(x)
|
||||
#endif
|
||||
|
||||
/* Assumes debugfs is mounted */
|
||||
const char *data_file = "/sys/kernel/tracing/user_events_data";
|
||||
const char *status_file = "/sys/kernel/tracing/user_events_status";
|
||||
int enabled = 0;
|
||||
|
||||
static int event_status(long **status)
|
||||
{
|
||||
int fd = open(status_file, O_RDONLY);
|
||||
|
||||
*status = mmap(NULL, sysconf(_SC_PAGESIZE), PROT_READ,
|
||||
MAP_SHARED, fd, 0);
|
||||
|
||||
close(fd);
|
||||
|
||||
if (*status == MAP_FAILED)
|
||||
return -1;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int event_reg(int fd, const char *command, long *index, long *mask,
|
||||
int *write)
|
||||
static int event_reg(int fd, const char *command, int *write, int *enabled)
|
||||
{
|
||||
struct user_reg reg = {0};
|
||||
|
||||
reg.size = sizeof(reg);
|
||||
reg.enable_bit = 31;
|
||||
reg.enable_size = sizeof(*enabled);
|
||||
reg.enable_addr = (__u64)enabled;
|
||||
reg.name_args = (__u64)command;
|
||||
|
||||
if (ioctl(fd, DIAG_IOCSREG, ®) == -1)
|
||||
return -1;
|
||||
|
||||
*index = reg.status_bit / __BITS_PER_LONG;
|
||||
*mask = endian_swap(1L << (reg.status_bit % __BITS_PER_LONG));
|
||||
*write = reg.write_index;
|
||||
|
||||
return 0;
|
||||
@ -62,17 +39,12 @@ static int event_reg(int fd, const char *command, long *index, long *mask,
|
||||
int main(int argc, char **argv)
|
||||
{
|
||||
int data_fd, write;
|
||||
long index, mask;
|
||||
long *status_page;
|
||||
struct iovec io[2];
|
||||
__u32 count = 0;
|
||||
|
||||
if (event_status(&status_page) == -1)
|
||||
return errno;
|
||||
|
||||
data_fd = open(data_file, O_RDWR);
|
||||
|
||||
if (event_reg(data_fd, "test u32 count", &index, &mask, &write) == -1)
|
||||
if (event_reg(data_fd, "test u32 count", &write, &enabled) == -1)
|
||||
return errno;
|
||||
|
||||
/* Setup iovec */
|
||||
@ -80,13 +52,12 @@ int main(int argc, char **argv)
|
||||
io[0].iov_len = sizeof(write);
|
||||
io[1].iov_base = &count;
|
||||
io[1].iov_len = sizeof(count);
|
||||
|
||||
ask:
|
||||
printf("Press enter to check status...\n");
|
||||
getchar();
|
||||
|
||||
/* Check if anyone is listening */
|
||||
if (status_page[index] & mask) {
|
||||
if (enabled) {
|
||||
/* Yep, trace out our data */
|
||||
writev(data_fd, (const struct iovec *)io, 2);
|
||||
|
||||
|
@ -61,6 +61,7 @@ my @skip_abs = (
|
||||
'/proc/device-tree',
|
||||
'/proc/1/syscall',
|
||||
'/sys/firmware/devicetree',
|
||||
'/sys/kernel/tracing/trace_pipe',
|
||||
'/sys/kernel/debug/tracing/trace_pipe',
|
||||
'/sys/kernel/security/apparmor/revision');
|
||||
|
||||
|
@ -110,6 +110,7 @@ static ssize_t uwrite(void const *const buf, size_t const count)
|
||||
{
|
||||
size_t cnt = count;
|
||||
off_t idx = 0;
|
||||
void *p = NULL;
|
||||
|
||||
file_updated = 1;
|
||||
|
||||
@ -117,7 +118,10 @@ static ssize_t uwrite(void const *const buf, size_t const count)
|
||||
off_t aoffset = (file_ptr + count) - file_end;
|
||||
|
||||
if (aoffset > file_append_size) {
|
||||
file_append = realloc(file_append, aoffset);
|
||||
p = realloc(file_append, aoffset);
|
||||
if (!p)
|
||||
free(file_append);
|
||||
file_append = p;
|
||||
file_append_size = aoffset;
|
||||
}
|
||||
if (!file_append) {
|
||||
|
@ -627,7 +627,7 @@ class TracepointProvider(Provider):
|
||||
name)'.
|
||||
|
||||
All available events have directories under
|
||||
/sys/kernel/debug/tracing/events/ which export information
|
||||
/sys/kernel/tracing/events/ which export information
|
||||
about the specific event. Therefore, listing the dirs gives us
|
||||
a list of all available events.
|
||||
|
||||
|
@ -98,7 +98,7 @@ int tracing_root_ok(void)
|
||||
void tracing_on(void)
|
||||
{
|
||||
#if CONTROL_TRACING > 0
|
||||
#define TRACEDIR "/sys/kernel/debug/tracing"
|
||||
#define TRACEDIR "/sys/kernel/tracing"
|
||||
char pidstr[32];
|
||||
|
||||
if (!tracing_root_ok())
|
||||
@ -124,7 +124,7 @@ void tracing_off(void)
|
||||
#if CONTROL_TRACING > 0
|
||||
if (!tracing_root_ok())
|
||||
return;
|
||||
cat_into_file("0", "/sys/kernel/debug/tracing/tracing_on");
|
||||
cat_into_file("0", "/sys/kernel/tracing/tracing_on");
|
||||
#endif
|
||||
}
|
||||
|
||||
|
@ -10,7 +10,7 @@ LDLIBS += -lrt -lpthread -lm
|
||||
# This test will not compile until user_events.h is added
|
||||
# back to uapi.
|
||||
|
||||
TEST_GEN_PROGS = ftrace_test dyn_test perf_test
|
||||
TEST_GEN_PROGS = ftrace_test dyn_test perf_test abi_test
|
||||
|
||||
TEST_FILES := settings
|
||||
|
||||
|
229
tools/testing/selftests/user_events/abi_test.c
Normal file
229
tools/testing/selftests/user_events/abi_test.c
Normal file
@ -0,0 +1,229 @@
|
||||
// SPDX-License-Identifier: GPL-2.0
|
||||
/*
|
||||
* User Events ABI Test Program
|
||||
*
|
||||
* Copyright (c) 2022 Beau Belgrave <beaub@linux.microsoft.com>
|
||||
*/
|
||||
|
||||
#define _GNU_SOURCE
|
||||
#include <sched.h>
|
||||
|
||||
#include <errno.h>
|
||||
#include <linux/user_events.h>
|
||||
#include <stdio.h>
|
||||
#include <stdlib.h>
|
||||
#include <fcntl.h>
|
||||
#include <sys/ioctl.h>
|
||||
#include <sys/stat.h>
|
||||
#include <unistd.h>
|
||||
#include <asm/unistd.h>
|
||||
|
||||
#include "../kselftest_harness.h"
|
||||
|
||||
const char *data_file = "/sys/kernel/tracing/user_events_data";
|
||||
const char *enable_file = "/sys/kernel/tracing/events/user_events/__abi_event/enable";
|
||||
|
||||
static int change_event(bool enable)
|
||||
{
|
||||
int fd = open(enable_file, O_RDWR);
|
||||
int ret;
|
||||
|
||||
if (fd < 0)
|
||||
return -1;
|
||||
|
||||
if (enable)
|
||||
ret = write(fd, "1", 1);
|
||||
else
|
||||
ret = write(fd, "0", 1);
|
||||
|
||||
close(fd);
|
||||
|
||||
if (ret == 1)
|
||||
ret = 0;
|
||||
else
|
||||
ret = -1;
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int reg_enable(long *enable, int size, int bit)
|
||||
{
|
||||
struct user_reg reg = {0};
|
||||
int fd = open(data_file, O_RDWR);
|
||||
int ret;
|
||||
|
||||
if (fd < 0)
|
||||
return -1;
|
||||
|
||||
reg.size = sizeof(reg);
|
||||
reg.name_args = (__u64)"__abi_event";
|
||||
reg.enable_bit = bit;
|
||||
reg.enable_addr = (__u64)enable;
|
||||
reg.enable_size = size;
|
||||
|
||||
ret = ioctl(fd, DIAG_IOCSREG, ®);
|
||||
|
||||
close(fd);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int reg_disable(long *enable, int bit)
|
||||
{
|
||||
struct user_unreg reg = {0};
|
||||
int fd = open(data_file, O_RDWR);
|
||||
int ret;
|
||||
|
||||
if (fd < 0)
|
||||
return -1;
|
||||
|
||||
reg.size = sizeof(reg);
|
||||
reg.disable_bit = bit;
|
||||
reg.disable_addr = (__u64)enable;
|
||||
|
||||
ret = ioctl(fd, DIAG_IOCSUNREG, ®);
|
||||
|
||||
close(fd);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
FIXTURE(user) {
|
||||
long check;
|
||||
};
|
||||
|
||||
FIXTURE_SETUP(user) {
|
||||
change_event(false);
|
||||
self->check = 0;
|
||||
}
|
||||
|
||||
FIXTURE_TEARDOWN(user) {
|
||||
}
|
||||
|
||||
TEST_F(user, enablement) {
|
||||
/* Changes should be reflected immediately */
|
||||
ASSERT_EQ(0, self->check);
|
||||
ASSERT_EQ(0, reg_enable(&self->check, sizeof(int), 0));
|
||||
ASSERT_EQ(0, change_event(true));
|
||||
ASSERT_EQ(1, self->check);
|
||||
ASSERT_EQ(0, change_event(false));
|
||||
ASSERT_EQ(0, self->check);
|
||||
|
||||
/* Ensure kernel clears bit after disable */
|
||||
ASSERT_EQ(0, change_event(true));
|
||||
ASSERT_EQ(1, self->check);
|
||||
ASSERT_EQ(0, reg_disable(&self->check, 0));
|
||||
ASSERT_EQ(0, self->check);
|
||||
|
||||
/* Ensure doesn't change after unreg */
|
||||
ASSERT_EQ(0, change_event(true));
|
||||
ASSERT_EQ(0, self->check);
|
||||
ASSERT_EQ(0, change_event(false));
|
||||
}
|
||||
|
||||
TEST_F(user, bit_sizes) {
|
||||
/* Allow 0-31 bits for 32-bit */
|
||||
ASSERT_EQ(0, reg_enable(&self->check, sizeof(int), 0));
|
||||
ASSERT_EQ(0, reg_enable(&self->check, sizeof(int), 31));
|
||||
ASSERT_NE(0, reg_enable(&self->check, sizeof(int), 32));
|
||||
ASSERT_EQ(0, reg_disable(&self->check, 0));
|
||||
ASSERT_EQ(0, reg_disable(&self->check, 31));
|
||||
|
||||
#if BITS_PER_LONG == 8
|
||||
/* Allow 0-64 bits for 64-bit */
|
||||
ASSERT_EQ(0, reg_enable(&self->check, sizeof(long), 63));
|
||||
ASSERT_NE(0, reg_enable(&self->check, sizeof(long), 64));
|
||||
ASSERT_EQ(0, reg_disable(&self->check, 63));
|
||||
#endif
|
||||
|
||||
/* Disallowed sizes (everything beside 4 and 8) */
|
||||
ASSERT_NE(0, reg_enable(&self->check, 1, 0));
|
||||
ASSERT_NE(0, reg_enable(&self->check, 2, 0));
|
||||
ASSERT_NE(0, reg_enable(&self->check, 3, 0));
|
||||
ASSERT_NE(0, reg_enable(&self->check, 5, 0));
|
||||
ASSERT_NE(0, reg_enable(&self->check, 6, 0));
|
||||
ASSERT_NE(0, reg_enable(&self->check, 7, 0));
|
||||
ASSERT_NE(0, reg_enable(&self->check, 9, 0));
|
||||
ASSERT_NE(0, reg_enable(&self->check, 128, 0));
|
||||
}
|
||||
|
||||
TEST_F(user, forks) {
|
||||
int i;
|
||||
|
||||
/* Ensure COW pages get updated after fork */
|
||||
ASSERT_EQ(0, reg_enable(&self->check, sizeof(int), 0));
|
||||
ASSERT_EQ(0, self->check);
|
||||
|
||||
if (fork() == 0) {
|
||||
/* Force COW */
|
||||
self->check = 0;
|
||||
|
||||
/* Up to 1 sec for enablement */
|
||||
for (i = 0; i < 10; ++i) {
|
||||
usleep(100000);
|
||||
|
||||
if (self->check)
|
||||
exit(0);
|
||||
}
|
||||
|
||||
exit(1);
|
||||
}
|
||||
|
||||
/* Allow generous time for COW, then enable */
|
||||
usleep(100000);
|
||||
ASSERT_EQ(0, change_event(true));
|
||||
|
||||
ASSERT_NE(-1, wait(&i));
|
||||
ASSERT_EQ(0, WEXITSTATUS(i));
|
||||
|
||||
/* Ensure child doesn't disable parent */
|
||||
if (fork() == 0)
|
||||
exit(reg_disable(&self->check, 0));
|
||||
|
||||
ASSERT_NE(-1, wait(&i));
|
||||
ASSERT_EQ(0, WEXITSTATUS(i));
|
||||
ASSERT_EQ(1, self->check);
|
||||
ASSERT_EQ(0, change_event(false));
|
||||
ASSERT_EQ(0, self->check);
|
||||
}
|
||||
|
||||
/* Waits up to 1 sec for enablement */
|
||||
static int clone_check(void *check)
|
||||
{
|
||||
int i;
|
||||
|
||||
for (i = 0; i < 10; ++i) {
|
||||
usleep(100000);
|
||||
|
||||
if (*(long *)check)
|
||||
return 0;
|
||||
}
|
||||
|
||||
return 1;
|
||||
}
|
||||
|
||||
TEST_F(user, clones) {
|
||||
int i, stack_size = 4096;
|
||||
void *stack = mmap(NULL, stack_size, PROT_READ | PROT_WRITE,
|
||||
MAP_PRIVATE | MAP_ANONYMOUS | MAP_STACK,
|
||||
-1, 0);
|
||||
|
||||
ASSERT_NE(MAP_FAILED, stack);
|
||||
ASSERT_EQ(0, reg_enable(&self->check, sizeof(int), 0));
|
||||
ASSERT_EQ(0, self->check);
|
||||
|
||||
/* Shared VM should see enablements */
|
||||
ASSERT_NE(-1, clone(&clone_check, stack + stack_size,
|
||||
CLONE_VM | SIGCHLD, &self->check));
|
||||
|
||||
ASSERT_EQ(0, change_event(true));
|
||||
ASSERT_NE(-1, wait(&i));
|
||||
ASSERT_EQ(0, WEXITSTATUS(i));
|
||||
munmap(stack, stack_size);
|
||||
ASSERT_EQ(0, change_event(false));
|
||||
}
|
||||
|
||||
int main(int argc, char **argv)
|
||||
{
|
||||
return test_harness_run(argc, argv);
|
||||
}
|
@ -16,7 +16,7 @@
|
||||
|
||||
#include "../kselftest_harness.h"
|
||||
|
||||
const char *dyn_file = "/sys/kernel/debug/tracing/dynamic_events";
|
||||
const char *dyn_file = "/sys/kernel/tracing/dynamic_events";
|
||||
const char *clear = "!u:__test_event";
|
||||
|
||||
static int Append(const char *value)
|
||||
|
@ -12,20 +12,16 @@
|
||||
#include <fcntl.h>
|
||||
#include <sys/ioctl.h>
|
||||
#include <sys/stat.h>
|
||||
#include <sys/uio.h>
|
||||
#include <unistd.h>
|
||||
|
||||
#include "../kselftest_harness.h"
|
||||
|
||||
const char *data_file = "/sys/kernel/debug/tracing/user_events_data";
|
||||
const char *status_file = "/sys/kernel/debug/tracing/user_events_status";
|
||||
const char *enable_file = "/sys/kernel/debug/tracing/events/user_events/__test_event/enable";
|
||||
const char *trace_file = "/sys/kernel/debug/tracing/trace";
|
||||
const char *fmt_file = "/sys/kernel/debug/tracing/events/user_events/__test_event/format";
|
||||
|
||||
static inline int status_check(char *status_page, int status_bit)
|
||||
{
|
||||
return status_page[status_bit >> 3] & (1 << (status_bit & 7));
|
||||
}
|
||||
const char *data_file = "/sys/kernel/tracing/user_events_data";
|
||||
const char *status_file = "/sys/kernel/tracing/user_events_status";
|
||||
const char *enable_file = "/sys/kernel/tracing/events/user_events/__test_event/enable";
|
||||
const char *trace_file = "/sys/kernel/tracing/trace";
|
||||
const char *fmt_file = "/sys/kernel/tracing/events/user_events/__test_event/format";
|
||||
|
||||
static int trace_bytes(void)
|
||||
{
|
||||
@ -106,13 +102,23 @@ err:
|
||||
return -1;
|
||||
}
|
||||
|
||||
static int clear(void)
|
||||
static int clear(int *check)
|
||||
{
|
||||
struct user_unreg unreg = {0};
|
||||
|
||||
unreg.size = sizeof(unreg);
|
||||
unreg.disable_bit = 31;
|
||||
unreg.disable_addr = (__u64)check;
|
||||
|
||||
int fd = open(data_file, O_RDWR);
|
||||
|
||||
if (fd == -1)
|
||||
return -1;
|
||||
|
||||
if (ioctl(fd, DIAG_IOCSUNREG, &unreg) == -1)
|
||||
if (errno != ENOENT)
|
||||
return -1;
|
||||
|
||||
if (ioctl(fd, DIAG_IOCSDEL, "__test_event") == -1)
|
||||
if (errno != ENOENT)
|
||||
return -1;
|
||||
@ -122,7 +128,7 @@ static int clear(void)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int check_print_fmt(const char *event, const char *expected)
|
||||
static int check_print_fmt(const char *event, const char *expected, int *check)
|
||||
{
|
||||
struct user_reg reg = {0};
|
||||
char print_fmt[256];
|
||||
@ -130,7 +136,7 @@ static int check_print_fmt(const char *event, const char *expected)
|
||||
int fd;
|
||||
|
||||
/* Ensure cleared */
|
||||
ret = clear();
|
||||
ret = clear(check);
|
||||
|
||||
if (ret != 0)
|
||||
return ret;
|
||||
@ -142,14 +148,19 @@ static int check_print_fmt(const char *event, const char *expected)
|
||||
|
||||
reg.size = sizeof(reg);
|
||||
reg.name_args = (__u64)event;
|
||||
reg.enable_bit = 31;
|
||||
reg.enable_addr = (__u64)check;
|
||||
reg.enable_size = sizeof(*check);
|
||||
|
||||
/* Register should work */
|
||||
ret = ioctl(fd, DIAG_IOCSREG, ®);
|
||||
|
||||
close(fd);
|
||||
|
||||
if (ret != 0)
|
||||
if (ret != 0) {
|
||||
printf("Reg failed in fmt\n");
|
||||
return ret;
|
||||
}
|
||||
|
||||
/* Ensure correct print_fmt */
|
||||
ret = get_print_fmt(print_fmt, sizeof(print_fmt));
|
||||
@ -164,6 +175,7 @@ FIXTURE(user) {
|
||||
int status_fd;
|
||||
int data_fd;
|
||||
int enable_fd;
|
||||
int check;
|
||||
};
|
||||
|
||||
FIXTURE_SETUP(user) {
|
||||
@ -185,59 +197,63 @@ FIXTURE_TEARDOWN(user) {
|
||||
close(self->enable_fd);
|
||||
}
|
||||
|
||||
ASSERT_EQ(0, clear());
|
||||
if (clear(&self->check) != 0)
|
||||
printf("WARNING: Clear didn't work!\n");
|
||||
}
|
||||
|
||||
TEST_F(user, register_events) {
|
||||
struct user_reg reg = {0};
|
||||
int page_size = sysconf(_SC_PAGESIZE);
|
||||
char *status_page;
|
||||
struct user_unreg unreg = {0};
|
||||
|
||||
reg.size = sizeof(reg);
|
||||
reg.name_args = (__u64)"__test_event u32 field1; u32 field2";
|
||||
reg.enable_bit = 31;
|
||||
reg.enable_addr = (__u64)&self->check;
|
||||
reg.enable_size = sizeof(self->check);
|
||||
|
||||
status_page = mmap(NULL, page_size, PROT_READ, MAP_SHARED,
|
||||
self->status_fd, 0);
|
||||
unreg.size = sizeof(unreg);
|
||||
unreg.disable_bit = 31;
|
||||
unreg.disable_addr = (__u64)&self->check;
|
||||
|
||||
/* Register should work */
|
||||
ASSERT_EQ(0, ioctl(self->data_fd, DIAG_IOCSREG, ®));
|
||||
ASSERT_EQ(0, reg.write_index);
|
||||
ASSERT_NE(0, reg.status_bit);
|
||||
|
||||
/* Multiple registers should result in same index */
|
||||
/* Multiple registers to the same addr + bit should fail */
|
||||
ASSERT_EQ(-1, ioctl(self->data_fd, DIAG_IOCSREG, ®));
|
||||
ASSERT_EQ(EADDRINUSE, errno);
|
||||
|
||||
/* Multiple registers to same name should result in same index */
|
||||
reg.enable_bit = 30;
|
||||
ASSERT_EQ(0, ioctl(self->data_fd, DIAG_IOCSREG, ®));
|
||||
ASSERT_EQ(0, reg.write_index);
|
||||
ASSERT_NE(0, reg.status_bit);
|
||||
|
||||
/* Ensure disabled */
|
||||
self->enable_fd = open(enable_file, O_RDWR);
|
||||
ASSERT_NE(-1, self->enable_fd);
|
||||
ASSERT_NE(-1, write(self->enable_fd, "0", sizeof("0")))
|
||||
|
||||
/* MMAP should work and be zero'd */
|
||||
ASSERT_NE(MAP_FAILED, status_page);
|
||||
ASSERT_NE(NULL, status_page);
|
||||
ASSERT_EQ(0, status_check(status_page, reg.status_bit));
|
||||
|
||||
/* Enable event and ensure bits updated in status */
|
||||
ASSERT_NE(-1, write(self->enable_fd, "1", sizeof("1")))
|
||||
ASSERT_NE(0, status_check(status_page, reg.status_bit));
|
||||
ASSERT_EQ(1 << reg.enable_bit, self->check);
|
||||
|
||||
/* Disable event and ensure bits updated in status */
|
||||
ASSERT_NE(-1, write(self->enable_fd, "0", sizeof("0")))
|
||||
ASSERT_EQ(0, status_check(status_page, reg.status_bit));
|
||||
ASSERT_EQ(0, self->check);
|
||||
|
||||
/* File still open should return -EBUSY for delete */
|
||||
ASSERT_EQ(-1, ioctl(self->data_fd, DIAG_IOCSDEL, "__test_event"));
|
||||
ASSERT_EQ(EBUSY, errno);
|
||||
|
||||
/* Delete should work only after close */
|
||||
/* Unregister */
|
||||
ASSERT_EQ(0, ioctl(self->data_fd, DIAG_IOCSUNREG, &unreg));
|
||||
unreg.disable_bit = 30;
|
||||
ASSERT_EQ(0, ioctl(self->data_fd, DIAG_IOCSUNREG, &unreg));
|
||||
|
||||
/* Delete should work only after close and unregister */
|
||||
close(self->data_fd);
|
||||
self->data_fd = open(data_file, O_RDWR);
|
||||
ASSERT_EQ(0, ioctl(self->data_fd, DIAG_IOCSDEL, "__test_event"));
|
||||
|
||||
/* Unmap should work */
|
||||
ASSERT_EQ(0, munmap(status_page, page_size));
|
||||
}
|
||||
|
||||
TEST_F(user, write_events) {
|
||||
@ -245,11 +261,12 @@ TEST_F(user, write_events) {
|
||||
struct iovec io[3];
|
||||
__u32 field1, field2;
|
||||
int before = 0, after = 0;
|
||||
int page_size = sysconf(_SC_PAGESIZE);
|
||||
char *status_page;
|
||||
|
||||
reg.size = sizeof(reg);
|
||||
reg.name_args = (__u64)"__test_event u32 field1; u32 field2";
|
||||
reg.enable_bit = 31;
|
||||
reg.enable_addr = (__u64)&self->check;
|
||||
reg.enable_size = sizeof(self->check);
|
||||
|
||||
field1 = 1;
|
||||
field2 = 2;
|
||||
@ -261,18 +278,10 @@ TEST_F(user, write_events) {
|
||||
io[2].iov_base = &field2;
|
||||
io[2].iov_len = sizeof(field2);
|
||||
|
||||
status_page = mmap(NULL, page_size, PROT_READ, MAP_SHARED,
|
||||
self->status_fd, 0);
|
||||
|
||||
/* Register should work */
|
||||
ASSERT_EQ(0, ioctl(self->data_fd, DIAG_IOCSREG, ®));
|
||||
ASSERT_EQ(0, reg.write_index);
|
||||
ASSERT_NE(0, reg.status_bit);
|
||||
|
||||
/* MMAP should work and be zero'd */
|
||||
ASSERT_NE(MAP_FAILED, status_page);
|
||||
ASSERT_NE(NULL, status_page);
|
||||
ASSERT_EQ(0, status_check(status_page, reg.status_bit));
|
||||
ASSERT_EQ(0, self->check);
|
||||
|
||||
/* Write should fail on invalid slot with ENOENT */
|
||||
io[0].iov_base = &field2;
|
||||
@ -287,13 +296,18 @@ TEST_F(user, write_events) {
|
||||
ASSERT_NE(-1, write(self->enable_fd, "1", sizeof("1")))
|
||||
|
||||
/* Event should now be enabled */
|
||||
ASSERT_NE(0, status_check(status_page, reg.status_bit));
|
||||
ASSERT_NE(1 << reg.enable_bit, self->check);
|
||||
|
||||
/* Write should make it out to ftrace buffers */
|
||||
before = trace_bytes();
|
||||
ASSERT_NE(-1, writev(self->data_fd, (const struct iovec *)io, 3));
|
||||
after = trace_bytes();
|
||||
ASSERT_GT(after, before);
|
||||
|
||||
/* Negative index should fail with EINVAL */
|
||||
reg.write_index = -1;
|
||||
ASSERT_EQ(-1, writev(self->data_fd, (const struct iovec *)io, 3));
|
||||
ASSERT_EQ(EINVAL, errno);
|
||||
}
|
||||
|
||||
TEST_F(user, write_fault) {
|
||||
@ -304,6 +318,9 @@ TEST_F(user, write_fault) {
|
||||
|
||||
reg.size = sizeof(reg);
|
||||
reg.name_args = (__u64)"__test_event u64 anon";
|
||||
reg.enable_bit = 31;
|
||||
reg.enable_addr = (__u64)&self->check;
|
||||
reg.enable_size = sizeof(self->check);
|
||||
|
||||
anon = mmap(NULL, l, PROT_READ, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
|
||||
ASSERT_NE(MAP_FAILED, anon);
|
||||
@ -316,7 +333,6 @@ TEST_F(user, write_fault) {
|
||||
/* Register should work */
|
||||
ASSERT_EQ(0, ioctl(self->data_fd, DIAG_IOCSREG, ®));
|
||||
ASSERT_EQ(0, reg.write_index);
|
||||
ASSERT_NE(0, reg.status_bit);
|
||||
|
||||
/* Write should work normally */
|
||||
ASSERT_NE(-1, writev(self->data_fd, (const struct iovec *)io, 2));
|
||||
@ -333,24 +349,17 @@ TEST_F(user, write_validator) {
|
||||
int loc, bytes;
|
||||
char data[8];
|
||||
int before = 0, after = 0;
|
||||
int page_size = sysconf(_SC_PAGESIZE);
|
||||
char *status_page;
|
||||
|
||||
status_page = mmap(NULL, page_size, PROT_READ, MAP_SHARED,
|
||||
self->status_fd, 0);
|
||||
|
||||
reg.size = sizeof(reg);
|
||||
reg.name_args = (__u64)"__test_event __rel_loc char[] data";
|
||||
reg.enable_bit = 31;
|
||||
reg.enable_addr = (__u64)&self->check;
|
||||
reg.enable_size = sizeof(self->check);
|
||||
|
||||
/* Register should work */
|
||||
ASSERT_EQ(0, ioctl(self->data_fd, DIAG_IOCSREG, ®));
|
||||
ASSERT_EQ(0, reg.write_index);
|
||||
ASSERT_NE(0, reg.status_bit);
|
||||
|
||||
/* MMAP should work and be zero'd */
|
||||
ASSERT_NE(MAP_FAILED, status_page);
|
||||
ASSERT_NE(NULL, status_page);
|
||||
ASSERT_EQ(0, status_check(status_page, reg.status_bit));
|
||||
ASSERT_EQ(0, self->check);
|
||||
|
||||
io[0].iov_base = ®.write_index;
|
||||
io[0].iov_len = sizeof(reg.write_index);
|
||||
@ -369,7 +378,7 @@ TEST_F(user, write_validator) {
|
||||
ASSERT_NE(-1, write(self->enable_fd, "1", sizeof("1")))
|
||||
|
||||
/* Event should now be enabled */
|
||||
ASSERT_NE(0, status_check(status_page, reg.status_bit));
|
||||
ASSERT_EQ(1 << reg.enable_bit, self->check);
|
||||
|
||||
/* Full in-bounds write should work */
|
||||
before = trace_bytes();
|
||||
@ -409,71 +418,88 @@ TEST_F(user, print_fmt) {
|
||||
int ret;
|
||||
|
||||
ret = check_print_fmt("__test_event __rel_loc char[] data",
|
||||
"print fmt: \"data=%s\", __get_rel_str(data)");
|
||||
"print fmt: \"data=%s\", __get_rel_str(data)",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event __data_loc char[] data",
|
||||
"print fmt: \"data=%s\", __get_str(data)");
|
||||
"print fmt: \"data=%s\", __get_str(data)",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event s64 data",
|
||||
"print fmt: \"data=%lld\", REC->data");
|
||||
"print fmt: \"data=%lld\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event u64 data",
|
||||
"print fmt: \"data=%llu\", REC->data");
|
||||
"print fmt: \"data=%llu\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event s32 data",
|
||||
"print fmt: \"data=%d\", REC->data");
|
||||
"print fmt: \"data=%d\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event u32 data",
|
||||
"print fmt: \"data=%u\", REC->data");
|
||||
"print fmt: \"data=%u\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event int data",
|
||||
"print fmt: \"data=%d\", REC->data");
|
||||
"print fmt: \"data=%d\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event unsigned int data",
|
||||
"print fmt: \"data=%u\", REC->data");
|
||||
"print fmt: \"data=%u\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event s16 data",
|
||||
"print fmt: \"data=%d\", REC->data");
|
||||
"print fmt: \"data=%d\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event u16 data",
|
||||
"print fmt: \"data=%u\", REC->data");
|
||||
"print fmt: \"data=%u\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event short data",
|
||||
"print fmt: \"data=%d\", REC->data");
|
||||
"print fmt: \"data=%d\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event unsigned short data",
|
||||
"print fmt: \"data=%u\", REC->data");
|
||||
"print fmt: \"data=%u\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event s8 data",
|
||||
"print fmt: \"data=%d\", REC->data");
|
||||
"print fmt: \"data=%d\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event u8 data",
|
||||
"print fmt: \"data=%u\", REC->data");
|
||||
"print fmt: \"data=%u\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event char data",
|
||||
"print fmt: \"data=%d\", REC->data");
|
||||
"print fmt: \"data=%d\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event unsigned char data",
|
||||
"print fmt: \"data=%u\", REC->data");
|
||||
"print fmt: \"data=%u\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
|
||||
ret = check_print_fmt("__test_event char[4] data",
|
||||
"print fmt: \"data=%s\", REC->data");
|
||||
"print fmt: \"data=%s\", REC->data",
|
||||
&self->check);
|
||||
ASSERT_EQ(0, ret);
|
||||
}
|
||||
|
||||
|
@ -18,10 +18,9 @@
|
||||
|
||||
#include "../kselftest_harness.h"
|
||||
|
||||
const char *data_file = "/sys/kernel/debug/tracing/user_events_data";
|
||||
const char *status_file = "/sys/kernel/debug/tracing/user_events_status";
|
||||
const char *id_file = "/sys/kernel/debug/tracing/events/user_events/__test_event/id";
|
||||
const char *fmt_file = "/sys/kernel/debug/tracing/events/user_events/__test_event/format";
|
||||
const char *data_file = "/sys/kernel/tracing/user_events_data";
|
||||
const char *id_file = "/sys/kernel/tracing/events/user_events/__test_event/id";
|
||||
const char *fmt_file = "/sys/kernel/tracing/events/user_events/__test_event/format";
|
||||
|
||||
struct event {
|
||||
__u32 index;
|
||||
@ -35,11 +34,6 @@ static long perf_event_open(struct perf_event_attr *pe, pid_t pid,
|
||||
return syscall(__NR_perf_event_open, pe, pid, cpu, group_fd, flags);
|
||||
}
|
||||
|
||||
static inline int status_check(char *status_page, int status_bit)
|
||||
{
|
||||
return status_page[status_bit >> 3] & (1 << (status_bit & 7));
|
||||
}
|
||||
|
||||
static int get_id(void)
|
||||
{
|
||||
FILE *fp = fopen(id_file, "r");
|
||||
@ -88,45 +82,38 @@ static int get_offset(void)
|
||||
}
|
||||
|
||||
FIXTURE(user) {
|
||||
int status_fd;
|
||||
int data_fd;
|
||||
int check;
|
||||
};
|
||||
|
||||
FIXTURE_SETUP(user) {
|
||||
self->status_fd = open(status_file, O_RDONLY);
|
||||
ASSERT_NE(-1, self->status_fd);
|
||||
|
||||
self->data_fd = open(data_file, O_RDWR);
|
||||
ASSERT_NE(-1, self->data_fd);
|
||||
}
|
||||
|
||||
FIXTURE_TEARDOWN(user) {
|
||||
close(self->status_fd);
|
||||
close(self->data_fd);
|
||||
}
|
||||
|
||||
TEST_F(user, perf_write) {
|
||||
struct perf_event_attr pe = {0};
|
||||
struct user_reg reg = {0};
|
||||
int page_size = sysconf(_SC_PAGESIZE);
|
||||
char *status_page;
|
||||
struct event event;
|
||||
struct perf_event_mmap_page *perf_page;
|
||||
int page_size = sysconf(_SC_PAGESIZE);
|
||||
int id, fd, offset;
|
||||
__u32 *val;
|
||||
|
||||
reg.size = sizeof(reg);
|
||||
reg.name_args = (__u64)"__test_event u32 field1; u32 field2";
|
||||
|
||||
status_page = mmap(NULL, page_size, PROT_READ, MAP_SHARED,
|
||||
self->status_fd, 0);
|
||||
ASSERT_NE(MAP_FAILED, status_page);
|
||||
reg.enable_bit = 31;
|
||||
reg.enable_addr = (__u64)&self->check;
|
||||
reg.enable_size = sizeof(self->check);
|
||||
|
||||
/* Register should work */
|
||||
ASSERT_EQ(0, ioctl(self->data_fd, DIAG_IOCSREG, ®));
|
||||
ASSERT_EQ(0, reg.write_index);
|
||||
ASSERT_NE(0, reg.status_bit);
|
||||
ASSERT_EQ(0, status_check(status_page, reg.status_bit));
|
||||
ASSERT_EQ(0, self->check);
|
||||
|
||||
/* Id should be there */
|
||||
id = get_id();
|
||||
@ -149,7 +136,7 @@ TEST_F(user, perf_write) {
|
||||
ASSERT_NE(MAP_FAILED, perf_page);
|
||||
|
||||
/* Status should be updated */
|
||||
ASSERT_NE(0, status_check(status_page, reg.status_bit));
|
||||
ASSERT_EQ(1 << reg.enable_bit, self->check);
|
||||
|
||||
event.index = reg.write_index;
|
||||
event.field1 = 0xc001;
|
||||
@ -165,6 +152,12 @@ TEST_F(user, perf_write) {
|
||||
/* Ensure correct */
|
||||
ASSERT_EQ(event.field1, *val++);
|
||||
ASSERT_EQ(event.field2, *val++);
|
||||
|
||||
munmap(perf_page, page_size * 2);
|
||||
close(fd);
|
||||
|
||||
/* Status should be updated */
|
||||
ASSERT_EQ(0, self->check);
|
||||
}
|
||||
|
||||
int main(int argc, char **argv)
|
||||
|
Loading…
Reference in New Issue
Block a user