kmemleak: change some global variables to int

They don't have to be atomic_t, because they are simple boolean toggles.

Signed-off-by: Li Zefan <lizefan@huawei.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
Li Zefan 2014-04-03 14:46:29 -07:00 committed by Linus Torvalds
parent 5f3bf19aeb
commit 8910ae896c

View File

@ -192,15 +192,15 @@ static struct kmem_cache *object_cache;
static struct kmem_cache *scan_area_cache; static struct kmem_cache *scan_area_cache;
/* set if tracing memory operations is enabled */ /* set if tracing memory operations is enabled */
static atomic_t kmemleak_enabled = ATOMIC_INIT(0); static int kmemleak_enabled;
/* set in the late_initcall if there were no errors */ /* set in the late_initcall if there were no errors */
static atomic_t kmemleak_initialized = ATOMIC_INIT(0); static int kmemleak_initialized;
/* enables or disables early logging of the memory operations */ /* enables or disables early logging of the memory operations */
static atomic_t kmemleak_early_log = ATOMIC_INIT(1); static int kmemleak_early_log = 1;
/* set if a kmemleak warning was issued */ /* set if a kmemleak warning was issued */
static atomic_t kmemleak_warning = ATOMIC_INIT(0); static int kmemleak_warning;
/* set if a fatal kmemleak error has occurred */ /* set if a fatal kmemleak error has occurred */
static atomic_t kmemleak_error = ATOMIC_INIT(0); static int kmemleak_error;
/* minimum and maximum address that may be valid pointers */ /* minimum and maximum address that may be valid pointers */
static unsigned long min_addr = ULONG_MAX; static unsigned long min_addr = ULONG_MAX;
@ -268,7 +268,7 @@ static void kmemleak_disable(void);
#define kmemleak_warn(x...) do { \ #define kmemleak_warn(x...) do { \
pr_warning(x); \ pr_warning(x); \
dump_stack(); \ dump_stack(); \
atomic_set(&kmemleak_warning, 1); \ kmemleak_warning = 1; \
} while (0) } while (0)
/* /*
@ -806,7 +806,7 @@ static void __init log_early(int op_type, const void *ptr, size_t size,
unsigned long flags; unsigned long flags;
struct early_log *log; struct early_log *log;
if (atomic_read(&kmemleak_error)) { if (kmemleak_error) {
/* kmemleak stopped recording, just count the requests */ /* kmemleak stopped recording, just count the requests */
crt_early_log++; crt_early_log++;
return; return;
@ -841,7 +841,7 @@ static void early_alloc(struct early_log *log)
unsigned long flags; unsigned long flags;
int i; int i;
if (!atomic_read(&kmemleak_enabled) || !log->ptr || IS_ERR(log->ptr)) if (!kmemleak_enabled || !log->ptr || IS_ERR(log->ptr))
return; return;
/* /*
@ -894,9 +894,9 @@ void __ref kmemleak_alloc(const void *ptr, size_t size, int min_count,
{ {
pr_debug("%s(0x%p, %zu, %d)\n", __func__, ptr, size, min_count); pr_debug("%s(0x%p, %zu, %d)\n", __func__, ptr, size, min_count);
if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr)) if (kmemleak_enabled && ptr && !IS_ERR(ptr))
create_object((unsigned long)ptr, size, min_count, gfp); create_object((unsigned long)ptr, size, min_count, gfp);
else if (atomic_read(&kmemleak_early_log)) else if (kmemleak_early_log)
log_early(KMEMLEAK_ALLOC, ptr, size, min_count); log_early(KMEMLEAK_ALLOC, ptr, size, min_count);
} }
EXPORT_SYMBOL_GPL(kmemleak_alloc); EXPORT_SYMBOL_GPL(kmemleak_alloc);
@ -920,11 +920,11 @@ void __ref kmemleak_alloc_percpu(const void __percpu *ptr, size_t size)
* Percpu allocations are only scanned and not reported as leaks * Percpu allocations are only scanned and not reported as leaks
* (min_count is set to 0). * (min_count is set to 0).
*/ */
if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr)) if (kmemleak_enabled && ptr && !IS_ERR(ptr))
for_each_possible_cpu(cpu) for_each_possible_cpu(cpu)
create_object((unsigned long)per_cpu_ptr(ptr, cpu), create_object((unsigned long)per_cpu_ptr(ptr, cpu),
size, 0, GFP_KERNEL); size, 0, GFP_KERNEL);
else if (atomic_read(&kmemleak_early_log)) else if (kmemleak_early_log)
log_early(KMEMLEAK_ALLOC_PERCPU, ptr, size, 0); log_early(KMEMLEAK_ALLOC_PERCPU, ptr, size, 0);
} }
EXPORT_SYMBOL_GPL(kmemleak_alloc_percpu); EXPORT_SYMBOL_GPL(kmemleak_alloc_percpu);
@ -940,9 +940,9 @@ void __ref kmemleak_free(const void *ptr)
{ {
pr_debug("%s(0x%p)\n", __func__, ptr); pr_debug("%s(0x%p)\n", __func__, ptr);
if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr)) if (kmemleak_enabled && ptr && !IS_ERR(ptr))
delete_object_full((unsigned long)ptr); delete_object_full((unsigned long)ptr);
else if (atomic_read(&kmemleak_early_log)) else if (kmemleak_early_log)
log_early(KMEMLEAK_FREE, ptr, 0, 0); log_early(KMEMLEAK_FREE, ptr, 0, 0);
} }
EXPORT_SYMBOL_GPL(kmemleak_free); EXPORT_SYMBOL_GPL(kmemleak_free);
@ -960,9 +960,9 @@ void __ref kmemleak_free_part(const void *ptr, size_t size)
{ {
pr_debug("%s(0x%p)\n", __func__, ptr); pr_debug("%s(0x%p)\n", __func__, ptr);
if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr)) if (kmemleak_enabled && ptr && !IS_ERR(ptr))
delete_object_part((unsigned long)ptr, size); delete_object_part((unsigned long)ptr, size);
else if (atomic_read(&kmemleak_early_log)) else if (kmemleak_early_log)
log_early(KMEMLEAK_FREE_PART, ptr, size, 0); log_early(KMEMLEAK_FREE_PART, ptr, size, 0);
} }
EXPORT_SYMBOL_GPL(kmemleak_free_part); EXPORT_SYMBOL_GPL(kmemleak_free_part);
@ -980,11 +980,11 @@ void __ref kmemleak_free_percpu(const void __percpu *ptr)
pr_debug("%s(0x%p)\n", __func__, ptr); pr_debug("%s(0x%p)\n", __func__, ptr);
if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr)) if (kmemleak_enabled && ptr && !IS_ERR(ptr))
for_each_possible_cpu(cpu) for_each_possible_cpu(cpu)
delete_object_full((unsigned long)per_cpu_ptr(ptr, delete_object_full((unsigned long)per_cpu_ptr(ptr,
cpu)); cpu));
else if (atomic_read(&kmemleak_early_log)) else if (kmemleak_early_log)
log_early(KMEMLEAK_FREE_PERCPU, ptr, 0, 0); log_early(KMEMLEAK_FREE_PERCPU, ptr, 0, 0);
} }
EXPORT_SYMBOL_GPL(kmemleak_free_percpu); EXPORT_SYMBOL_GPL(kmemleak_free_percpu);
@ -1000,9 +1000,9 @@ void __ref kmemleak_not_leak(const void *ptr)
{ {
pr_debug("%s(0x%p)\n", __func__, ptr); pr_debug("%s(0x%p)\n", __func__, ptr);
if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr)) if (kmemleak_enabled && ptr && !IS_ERR(ptr))
make_gray_object((unsigned long)ptr); make_gray_object((unsigned long)ptr);
else if (atomic_read(&kmemleak_early_log)) else if (kmemleak_early_log)
log_early(KMEMLEAK_NOT_LEAK, ptr, 0, 0); log_early(KMEMLEAK_NOT_LEAK, ptr, 0, 0);
} }
EXPORT_SYMBOL(kmemleak_not_leak); EXPORT_SYMBOL(kmemleak_not_leak);
@ -1020,9 +1020,9 @@ void __ref kmemleak_ignore(const void *ptr)
{ {
pr_debug("%s(0x%p)\n", __func__, ptr); pr_debug("%s(0x%p)\n", __func__, ptr);
if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr)) if (kmemleak_enabled && ptr && !IS_ERR(ptr))
make_black_object((unsigned long)ptr); make_black_object((unsigned long)ptr);
else if (atomic_read(&kmemleak_early_log)) else if (kmemleak_early_log)
log_early(KMEMLEAK_IGNORE, ptr, 0, 0); log_early(KMEMLEAK_IGNORE, ptr, 0, 0);
} }
EXPORT_SYMBOL(kmemleak_ignore); EXPORT_SYMBOL(kmemleak_ignore);
@ -1042,9 +1042,9 @@ void __ref kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp)
{ {
pr_debug("%s(0x%p)\n", __func__, ptr); pr_debug("%s(0x%p)\n", __func__, ptr);
if (atomic_read(&kmemleak_enabled) && ptr && size && !IS_ERR(ptr)) if (kmemleak_enabled && ptr && size && !IS_ERR(ptr))
add_scan_area((unsigned long)ptr, size, gfp); add_scan_area((unsigned long)ptr, size, gfp);
else if (atomic_read(&kmemleak_early_log)) else if (kmemleak_early_log)
log_early(KMEMLEAK_SCAN_AREA, ptr, size, 0); log_early(KMEMLEAK_SCAN_AREA, ptr, size, 0);
} }
EXPORT_SYMBOL(kmemleak_scan_area); EXPORT_SYMBOL(kmemleak_scan_area);
@ -1062,9 +1062,9 @@ void __ref kmemleak_no_scan(const void *ptr)
{ {
pr_debug("%s(0x%p)\n", __func__, ptr); pr_debug("%s(0x%p)\n", __func__, ptr);
if (atomic_read(&kmemleak_enabled) && ptr && !IS_ERR(ptr)) if (kmemleak_enabled && ptr && !IS_ERR(ptr))
object_no_scan((unsigned long)ptr); object_no_scan((unsigned long)ptr);
else if (atomic_read(&kmemleak_early_log)) else if (kmemleak_early_log)
log_early(KMEMLEAK_NO_SCAN, ptr, 0, 0); log_early(KMEMLEAK_NO_SCAN, ptr, 0, 0);
} }
EXPORT_SYMBOL(kmemleak_no_scan); EXPORT_SYMBOL(kmemleak_no_scan);
@ -1089,7 +1089,7 @@ static bool update_checksum(struct kmemleak_object *object)
*/ */
static int scan_should_stop(void) static int scan_should_stop(void)
{ {
if (!atomic_read(&kmemleak_enabled)) if (!kmemleak_enabled)
return 1; return 1;
/* /*
@ -1630,14 +1630,14 @@ static ssize_t kmemleak_write(struct file *file, const char __user *user_buf,
return ret; return ret;
if (strncmp(buf, "clear", 5) == 0) { if (strncmp(buf, "clear", 5) == 0) {
if (atomic_read(&kmemleak_enabled)) if (kmemleak_enabled)
kmemleak_clear(); kmemleak_clear();
else else
__kmemleak_do_cleanup(); __kmemleak_do_cleanup();
goto out; goto out;
} }
if (!atomic_read(&kmemleak_enabled)) { if (!kmemleak_enabled) {
ret = -EBUSY; ret = -EBUSY;
goto out; goto out;
} }
@ -1726,14 +1726,14 @@ static DECLARE_WORK(cleanup_work, kmemleak_do_cleanup);
static void kmemleak_disable(void) static void kmemleak_disable(void)
{ {
/* atomically check whether it was already invoked */ /* atomically check whether it was already invoked */
if (atomic_cmpxchg(&kmemleak_error, 0, 1)) if (cmpxchg(&kmemleak_error, 0, 1))
return; return;
/* stop any memory operation tracing */ /* stop any memory operation tracing */
atomic_set(&kmemleak_enabled, 0); kmemleak_enabled = 0;
/* check whether it is too early for a kernel thread */ /* check whether it is too early for a kernel thread */
if (atomic_read(&kmemleak_initialized)) if (kmemleak_initialized)
schedule_work(&cleanup_work); schedule_work(&cleanup_work);
pr_info("Kernel memory leak detector disabled\n"); pr_info("Kernel memory leak detector disabled\n");
@ -1775,9 +1775,10 @@ void __init kmemleak_init(void)
int i; int i;
unsigned long flags; unsigned long flags;
kmemleak_early_log = 0;
#ifdef CONFIG_DEBUG_KMEMLEAK_DEFAULT_OFF #ifdef CONFIG_DEBUG_KMEMLEAK_DEFAULT_OFF
if (!kmemleak_skip_disable) { if (!kmemleak_skip_disable) {
atomic_set(&kmemleak_early_log, 0);
kmemleak_disable(); kmemleak_disable();
return; return;
} }
@ -1795,12 +1796,11 @@ void __init kmemleak_init(void)
/* the kernel is still in UP mode, so disabling the IRQs is enough */ /* the kernel is still in UP mode, so disabling the IRQs is enough */
local_irq_save(flags); local_irq_save(flags);
atomic_set(&kmemleak_early_log, 0); if (kmemleak_error) {
if (atomic_read(&kmemleak_error)) {
local_irq_restore(flags); local_irq_restore(flags);
return; return;
} else } else
atomic_set(&kmemleak_enabled, 1); kmemleak_enabled = 1;
local_irq_restore(flags); local_irq_restore(flags);
/* /*
@ -1844,9 +1844,9 @@ void __init kmemleak_init(void)
log->op_type); log->op_type);
} }
if (atomic_read(&kmemleak_warning)) { if (kmemleak_warning) {
print_log_trace(log); print_log_trace(log);
atomic_set(&kmemleak_warning, 0); kmemleak_warning = 0;
} }
} }
} }
@ -1858,9 +1858,9 @@ static int __init kmemleak_late_init(void)
{ {
struct dentry *dentry; struct dentry *dentry;
atomic_set(&kmemleak_initialized, 1); kmemleak_initialized = 1;
if (atomic_read(&kmemleak_error)) { if (kmemleak_error) {
/* /*
* Some error occurred and kmemleak was disabled. There is a * Some error occurred and kmemleak was disabled. There is a
* small chance that kmemleak_disable() was called immediately * small chance that kmemleak_disable() was called immediately