mirror of
https://github.com/torvalds/linux.git
synced 2024-12-27 05:11:48 +00:00
mm: account pmd page tables to the process
Dave noticed that unprivileged process can allocate significant amount of memory -- >500 MiB on x86_64 -- and stay unnoticed by oom-killer and memory cgroup. The trick is to allocate a lot of PMD page tables. Linux kernel doesn't account PMD tables to the process, only PTE. The use-cases below use few tricks to allocate a lot of PMD page tables while keeping VmRSS and VmPTE low. oom_score for the process will be 0. #include <errno.h> #include <stdio.h> #include <stdlib.h> #include <unistd.h> #include <sys/mman.h> #include <sys/prctl.h> #define PUD_SIZE (1UL << 30) #define PMD_SIZE (1UL << 21) #define NR_PUD 130000 int main(void) { char *addr = NULL; unsigned long i; prctl(PR_SET_THP_DISABLE); for (i = 0; i < NR_PUD ; i++) { addr = mmap(addr + PUD_SIZE, PUD_SIZE, PROT_WRITE|PROT_READ, MAP_ANONYMOUS|MAP_PRIVATE, -1, 0); if (addr == MAP_FAILED) { perror("mmap"); break; } *addr = 'x'; munmap(addr, PMD_SIZE); mmap(addr, PMD_SIZE, PROT_WRITE|PROT_READ, MAP_ANONYMOUS|MAP_PRIVATE|MAP_FIXED, -1, 0); if (addr == MAP_FAILED) perror("re-mmap"), exit(1); } printf("PID %d consumed %lu KiB in PMD page tables\n", getpid(), i * 4096 >> 10); return pause(); } The patch addresses the issue by account PMD tables to the process the same way we account PTE. The main place where PMD tables is accounted is __pmd_alloc() and free_pmd_range(). But there're few corner cases: - HugeTLB can share PMD page tables. The patch handles by accounting the table to all processes who share it. - x86 PAE pre-allocates few PMD tables on fork. - Architectures with FIRST_USER_ADDRESS > 0. We need to adjust sanity check on exit(2). Accounting only happens on configuration where PMD page table's level is present (PMD is not folded). As with nr_ptes we use per-mm counter. The counter value is used to calculate baseline for badness score by oom-killer. Signed-off-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com> Reported-by: Dave Hansen <dave.hansen@linux.intel.com> Cc: Hugh Dickins <hughd@google.com> Reviewed-by: Cyrill Gorcunov <gorcunov@openvz.org> Cc: Pavel Emelyanov <xemul@openvz.org> Cc: David Rientjes <rientjes@google.com> Tested-by: Sedat Dilek <sedat.dilek@gmail.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
8aa76875dc
commit
dc6c9a35b6
@ -555,12 +555,12 @@ this is causing problems for your system/application.
|
|||||||
|
|
||||||
oom_dump_tasks
|
oom_dump_tasks
|
||||||
|
|
||||||
Enables a system-wide task dump (excluding kernel threads) to be
|
Enables a system-wide task dump (excluding kernel threads) to be produced
|
||||||
produced when the kernel performs an OOM-killing and includes such
|
when the kernel performs an OOM-killing and includes such information as
|
||||||
information as pid, uid, tgid, vm size, rss, nr_ptes, swapents,
|
pid, uid, tgid, vm size, rss, nr_ptes, nr_pmds, swapents, oom_score_adj
|
||||||
oom_score_adj score, and name. This is helpful to determine why the
|
score, and name. This is helpful to determine why the OOM killer was
|
||||||
OOM killer was invoked, to identify the rogue task that caused it,
|
invoked, to identify the rogue task that caused it, and to determine why
|
||||||
and to determine why the OOM killer chose the task it did to kill.
|
the OOM killer chose the task it did to kill.
|
||||||
|
|
||||||
If this is set to zero, this information is suppressed. On very
|
If this is set to zero, this information is suppressed. On very
|
||||||
large systems with thousands of tasks it may not be feasible to dump
|
large systems with thousands of tasks it may not be feasible to dump
|
||||||
|
@ -190,7 +190,7 @@ void pud_populate(struct mm_struct *mm, pud_t *pudp, pmd_t *pmd)
|
|||||||
|
|
||||||
#endif /* CONFIG_X86_PAE */
|
#endif /* CONFIG_X86_PAE */
|
||||||
|
|
||||||
static void free_pmds(pmd_t *pmds[])
|
static void free_pmds(struct mm_struct *mm, pmd_t *pmds[])
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
@ -198,10 +198,11 @@ static void free_pmds(pmd_t *pmds[])
|
|||||||
if (pmds[i]) {
|
if (pmds[i]) {
|
||||||
pgtable_pmd_page_dtor(virt_to_page(pmds[i]));
|
pgtable_pmd_page_dtor(virt_to_page(pmds[i]));
|
||||||
free_page((unsigned long)pmds[i]);
|
free_page((unsigned long)pmds[i]);
|
||||||
|
mm_dec_nr_pmds(mm);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static int preallocate_pmds(pmd_t *pmds[])
|
static int preallocate_pmds(struct mm_struct *mm, pmd_t *pmds[])
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
bool failed = false;
|
bool failed = false;
|
||||||
@ -215,11 +216,13 @@ static int preallocate_pmds(pmd_t *pmds[])
|
|||||||
pmd = NULL;
|
pmd = NULL;
|
||||||
failed = true;
|
failed = true;
|
||||||
}
|
}
|
||||||
|
if (pmd)
|
||||||
|
mm_inc_nr_pmds(mm);
|
||||||
pmds[i] = pmd;
|
pmds[i] = pmd;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (failed) {
|
if (failed) {
|
||||||
free_pmds(pmds);
|
free_pmds(mm, pmds);
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -246,6 +249,7 @@ static void pgd_mop_up_pmds(struct mm_struct *mm, pgd_t *pgdp)
|
|||||||
|
|
||||||
paravirt_release_pmd(pgd_val(pgd) >> PAGE_SHIFT);
|
paravirt_release_pmd(pgd_val(pgd) >> PAGE_SHIFT);
|
||||||
pmd_free(mm, pmd);
|
pmd_free(mm, pmd);
|
||||||
|
mm_dec_nr_pmds(mm);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -283,7 +287,7 @@ pgd_t *pgd_alloc(struct mm_struct *mm)
|
|||||||
|
|
||||||
mm->pgd = pgd;
|
mm->pgd = pgd;
|
||||||
|
|
||||||
if (preallocate_pmds(pmds) != 0)
|
if (preallocate_pmds(mm, pmds) != 0)
|
||||||
goto out_free_pgd;
|
goto out_free_pgd;
|
||||||
|
|
||||||
if (paravirt_pgd_alloc(mm) != 0)
|
if (paravirt_pgd_alloc(mm) != 0)
|
||||||
@ -304,7 +308,7 @@ pgd_t *pgd_alloc(struct mm_struct *mm)
|
|||||||
return pgd;
|
return pgd;
|
||||||
|
|
||||||
out_free_pmds:
|
out_free_pmds:
|
||||||
free_pmds(pmds);
|
free_pmds(mm, pmds);
|
||||||
out_free_pgd:
|
out_free_pgd:
|
||||||
free_page((unsigned long)pgd);
|
free_page((unsigned long)pgd);
|
||||||
out:
|
out:
|
||||||
|
@ -21,7 +21,7 @@
|
|||||||
|
|
||||||
void task_mem(struct seq_file *m, struct mm_struct *mm)
|
void task_mem(struct seq_file *m, struct mm_struct *mm)
|
||||||
{
|
{
|
||||||
unsigned long data, text, lib, swap;
|
unsigned long data, text, lib, swap, ptes, pmds;
|
||||||
unsigned long hiwater_vm, total_vm, hiwater_rss, total_rss;
|
unsigned long hiwater_vm, total_vm, hiwater_rss, total_rss;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -42,6 +42,8 @@ void task_mem(struct seq_file *m, struct mm_struct *mm)
|
|||||||
text = (PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK)) >> 10;
|
text = (PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK)) >> 10;
|
||||||
lib = (mm->exec_vm << (PAGE_SHIFT-10)) - text;
|
lib = (mm->exec_vm << (PAGE_SHIFT-10)) - text;
|
||||||
swap = get_mm_counter(mm, MM_SWAPENTS);
|
swap = get_mm_counter(mm, MM_SWAPENTS);
|
||||||
|
ptes = PTRS_PER_PTE * sizeof(pte_t) * atomic_long_read(&mm->nr_ptes);
|
||||||
|
pmds = PTRS_PER_PMD * sizeof(pmd_t) * mm_nr_pmds(mm);
|
||||||
seq_printf(m,
|
seq_printf(m,
|
||||||
"VmPeak:\t%8lu kB\n"
|
"VmPeak:\t%8lu kB\n"
|
||||||
"VmSize:\t%8lu kB\n"
|
"VmSize:\t%8lu kB\n"
|
||||||
@ -54,6 +56,7 @@ void task_mem(struct seq_file *m, struct mm_struct *mm)
|
|||||||
"VmExe:\t%8lu kB\n"
|
"VmExe:\t%8lu kB\n"
|
||||||
"VmLib:\t%8lu kB\n"
|
"VmLib:\t%8lu kB\n"
|
||||||
"VmPTE:\t%8lu kB\n"
|
"VmPTE:\t%8lu kB\n"
|
||||||
|
"VmPMD:\t%8lu kB\n"
|
||||||
"VmSwap:\t%8lu kB\n",
|
"VmSwap:\t%8lu kB\n",
|
||||||
hiwater_vm << (PAGE_SHIFT-10),
|
hiwater_vm << (PAGE_SHIFT-10),
|
||||||
total_vm << (PAGE_SHIFT-10),
|
total_vm << (PAGE_SHIFT-10),
|
||||||
@ -63,8 +66,8 @@ void task_mem(struct seq_file *m, struct mm_struct *mm)
|
|||||||
total_rss << (PAGE_SHIFT-10),
|
total_rss << (PAGE_SHIFT-10),
|
||||||
data << (PAGE_SHIFT-10),
|
data << (PAGE_SHIFT-10),
|
||||||
mm->stack_vm << (PAGE_SHIFT-10), text, lib,
|
mm->stack_vm << (PAGE_SHIFT-10), text, lib,
|
||||||
(PTRS_PER_PTE * sizeof(pte_t) *
|
ptes >> 10,
|
||||||
atomic_long_read(&mm->nr_ptes)) >> 10,
|
pmds >> 10,
|
||||||
swap << (PAGE_SHIFT-10));
|
swap << (PAGE_SHIFT-10));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1438,8 +1438,32 @@ static inline int __pmd_alloc(struct mm_struct *mm, pud_t *pud,
|
|||||||
{
|
{
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline unsigned long mm_nr_pmds(struct mm_struct *mm)
|
||||||
|
{
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline void mm_inc_nr_pmds(struct mm_struct *mm) {}
|
||||||
|
static inline void mm_dec_nr_pmds(struct mm_struct *mm) {}
|
||||||
|
|
||||||
#else
|
#else
|
||||||
int __pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address);
|
int __pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address);
|
||||||
|
|
||||||
|
static inline unsigned long mm_nr_pmds(struct mm_struct *mm)
|
||||||
|
{
|
||||||
|
return atomic_long_read(&mm->nr_pmds);
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline void mm_inc_nr_pmds(struct mm_struct *mm)
|
||||||
|
{
|
||||||
|
atomic_long_inc(&mm->nr_pmds);
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline void mm_dec_nr_pmds(struct mm_struct *mm)
|
||||||
|
{
|
||||||
|
atomic_long_dec(&mm->nr_pmds);
|
||||||
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
int __pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma,
|
int __pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma,
|
||||||
|
@ -363,7 +363,8 @@ struct mm_struct {
|
|||||||
pgd_t * pgd;
|
pgd_t * pgd;
|
||||||
atomic_t mm_users; /* How many users with user space? */
|
atomic_t mm_users; /* How many users with user space? */
|
||||||
atomic_t mm_count; /* How many references to "struct mm_struct" (users count as 1) */
|
atomic_t mm_count; /* How many references to "struct mm_struct" (users count as 1) */
|
||||||
atomic_long_t nr_ptes; /* Page table pages */
|
atomic_long_t nr_ptes; /* PTE page table pages */
|
||||||
|
atomic_long_t nr_pmds; /* PMD page table pages */
|
||||||
int map_count; /* number of VMAs */
|
int map_count; /* number of VMAs */
|
||||||
|
|
||||||
spinlock_t page_table_lock; /* Protects page tables and some counters */
|
spinlock_t page_table_lock; /* Protects page tables and some counters */
|
||||||
|
@ -555,6 +555,9 @@ static struct mm_struct *mm_init(struct mm_struct *mm, struct task_struct *p)
|
|||||||
INIT_LIST_HEAD(&mm->mmlist);
|
INIT_LIST_HEAD(&mm->mmlist);
|
||||||
mm->core_state = NULL;
|
mm->core_state = NULL;
|
||||||
atomic_long_set(&mm->nr_ptes, 0);
|
atomic_long_set(&mm->nr_ptes, 0);
|
||||||
|
#ifndef __PAGETABLE_PMD_FOLDED
|
||||||
|
atomic_long_set(&mm->nr_pmds, 0);
|
||||||
|
#endif
|
||||||
mm->map_count = 0;
|
mm->map_count = 0;
|
||||||
mm->locked_vm = 0;
|
mm->locked_vm = 0;
|
||||||
mm->pinned_vm = 0;
|
mm->pinned_vm = 0;
|
||||||
|
@ -173,7 +173,7 @@ void dump_mm(const struct mm_struct *mm)
|
|||||||
"get_unmapped_area %p\n"
|
"get_unmapped_area %p\n"
|
||||||
#endif
|
#endif
|
||||||
"mmap_base %lu mmap_legacy_base %lu highest_vm_end %lu\n"
|
"mmap_base %lu mmap_legacy_base %lu highest_vm_end %lu\n"
|
||||||
"pgd %p mm_users %d mm_count %d nr_ptes %lu map_count %d\n"
|
"pgd %p mm_users %d mm_count %d nr_ptes %lu nr_pmds %lu map_count %d\n"
|
||||||
"hiwater_rss %lx hiwater_vm %lx total_vm %lx locked_vm %lx\n"
|
"hiwater_rss %lx hiwater_vm %lx total_vm %lx locked_vm %lx\n"
|
||||||
"pinned_vm %lx shared_vm %lx exec_vm %lx stack_vm %lx\n"
|
"pinned_vm %lx shared_vm %lx exec_vm %lx stack_vm %lx\n"
|
||||||
"start_code %lx end_code %lx start_data %lx end_data %lx\n"
|
"start_code %lx end_code %lx start_data %lx end_data %lx\n"
|
||||||
@ -206,6 +206,7 @@ void dump_mm(const struct mm_struct *mm)
|
|||||||
mm->pgd, atomic_read(&mm->mm_users),
|
mm->pgd, atomic_read(&mm->mm_users),
|
||||||
atomic_read(&mm->mm_count),
|
atomic_read(&mm->mm_count),
|
||||||
atomic_long_read((atomic_long_t *)&mm->nr_ptes),
|
atomic_long_read((atomic_long_t *)&mm->nr_ptes),
|
||||||
|
mm_nr_pmds((struct mm_struct *)mm),
|
||||||
mm->map_count,
|
mm->map_count,
|
||||||
mm->hiwater_rss, mm->hiwater_vm, mm->total_vm, mm->locked_vm,
|
mm->hiwater_rss, mm->hiwater_vm, mm->total_vm, mm->locked_vm,
|
||||||
mm->pinned_vm, mm->shared_vm, mm->exec_vm, mm->stack_vm,
|
mm->pinned_vm, mm->shared_vm, mm->exec_vm, mm->stack_vm,
|
||||||
|
@ -3598,6 +3598,7 @@ pte_t *huge_pmd_share(struct mm_struct *mm, unsigned long addr, pud_t *pud)
|
|||||||
if (saddr) {
|
if (saddr) {
|
||||||
spte = huge_pte_offset(svma->vm_mm, saddr);
|
spte = huge_pte_offset(svma->vm_mm, saddr);
|
||||||
if (spte) {
|
if (spte) {
|
||||||
|
mm_inc_nr_pmds(mm);
|
||||||
get_page(virt_to_page(spte));
|
get_page(virt_to_page(spte));
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
@ -3609,11 +3610,13 @@ pte_t *huge_pmd_share(struct mm_struct *mm, unsigned long addr, pud_t *pud)
|
|||||||
|
|
||||||
ptl = huge_pte_lockptr(hstate_vma(vma), mm, spte);
|
ptl = huge_pte_lockptr(hstate_vma(vma), mm, spte);
|
||||||
spin_lock(ptl);
|
spin_lock(ptl);
|
||||||
if (pud_none(*pud))
|
if (pud_none(*pud)) {
|
||||||
pud_populate(mm, pud,
|
pud_populate(mm, pud,
|
||||||
(pmd_t *)((unsigned long)spte & PAGE_MASK));
|
(pmd_t *)((unsigned long)spte & PAGE_MASK));
|
||||||
else
|
} else {
|
||||||
put_page(virt_to_page(spte));
|
put_page(virt_to_page(spte));
|
||||||
|
mm_inc_nr_pmds(mm);
|
||||||
|
}
|
||||||
spin_unlock(ptl);
|
spin_unlock(ptl);
|
||||||
out:
|
out:
|
||||||
pte = (pte_t *)pmd_alloc(mm, pud, addr);
|
pte = (pte_t *)pmd_alloc(mm, pud, addr);
|
||||||
@ -3644,6 +3647,7 @@ int huge_pmd_unshare(struct mm_struct *mm, unsigned long *addr, pte_t *ptep)
|
|||||||
|
|
||||||
pud_clear(pud);
|
pud_clear(pud);
|
||||||
put_page(virt_to_page(ptep));
|
put_page(virt_to_page(ptep));
|
||||||
|
mm_dec_nr_pmds(mm);
|
||||||
*addr = ALIGN(*addr, HPAGE_SIZE * PTRS_PER_PTE) - HPAGE_SIZE;
|
*addr = ALIGN(*addr, HPAGE_SIZE * PTRS_PER_PTE) - HPAGE_SIZE;
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
15
mm/memory.c
15
mm/memory.c
@ -428,6 +428,7 @@ static inline void free_pmd_range(struct mmu_gather *tlb, pud_t *pud,
|
|||||||
pmd = pmd_offset(pud, start);
|
pmd = pmd_offset(pud, start);
|
||||||
pud_clear(pud);
|
pud_clear(pud);
|
||||||
pmd_free_tlb(tlb, pmd, start);
|
pmd_free_tlb(tlb, pmd, start);
|
||||||
|
mm_dec_nr_pmds(tlb->mm);
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void free_pud_range(struct mmu_gather *tlb, pgd_t *pgd,
|
static inline void free_pud_range(struct mmu_gather *tlb, pgd_t *pgd,
|
||||||
@ -3322,15 +3323,17 @@ int __pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address)
|
|||||||
|
|
||||||
spin_lock(&mm->page_table_lock);
|
spin_lock(&mm->page_table_lock);
|
||||||
#ifndef __ARCH_HAS_4LEVEL_HACK
|
#ifndef __ARCH_HAS_4LEVEL_HACK
|
||||||
if (pud_present(*pud)) /* Another has populated it */
|
if (!pud_present(*pud)) {
|
||||||
pmd_free(mm, new);
|
mm_inc_nr_pmds(mm);
|
||||||
else
|
|
||||||
pud_populate(mm, pud, new);
|
pud_populate(mm, pud, new);
|
||||||
#else
|
} else /* Another has populated it */
|
||||||
if (pgd_present(*pud)) /* Another has populated it */
|
|
||||||
pmd_free(mm, new);
|
pmd_free(mm, new);
|
||||||
else
|
#else
|
||||||
|
if (!pgd_present(*pud)) {
|
||||||
|
mm_inc_nr_pmds(mm);
|
||||||
pgd_populate(mm, pud, new);
|
pgd_populate(mm, pud, new);
|
||||||
|
} else /* Another has populated it */
|
||||||
|
pmd_free(mm, new);
|
||||||
#endif /* __ARCH_HAS_4LEVEL_HACK */
|
#endif /* __ARCH_HAS_4LEVEL_HACK */
|
||||||
spin_unlock(&mm->page_table_lock);
|
spin_unlock(&mm->page_table_lock);
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -2853,7 +2853,9 @@ void exit_mmap(struct mm_struct *mm)
|
|||||||
vm_unacct_memory(nr_accounted);
|
vm_unacct_memory(nr_accounted);
|
||||||
|
|
||||||
WARN_ON(atomic_long_read(&mm->nr_ptes) >
|
WARN_ON(atomic_long_read(&mm->nr_ptes) >
|
||||||
(FIRST_USER_ADDRESS+PMD_SIZE-1)>>PMD_SHIFT);
|
round_up(FIRST_USER_ADDRESS, PMD_SIZE) >> PMD_SHIFT);
|
||||||
|
WARN_ON(mm_nr_pmds(mm) >
|
||||||
|
round_up(FIRST_USER_ADDRESS, PUD_SIZE) >> PUD_SHIFT);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Insert vm structure into process list sorted by address
|
/* Insert vm structure into process list sorted by address
|
||||||
|
@ -169,8 +169,8 @@ unsigned long oom_badness(struct task_struct *p, struct mem_cgroup *memcg,
|
|||||||
* The baseline for the badness score is the proportion of RAM that each
|
* The baseline for the badness score is the proportion of RAM that each
|
||||||
* task's rss, pagetable and swap space use.
|
* task's rss, pagetable and swap space use.
|
||||||
*/
|
*/
|
||||||
points = get_mm_rss(p->mm) + atomic_long_read(&p->mm->nr_ptes) +
|
points = get_mm_rss(p->mm) + get_mm_counter(p->mm, MM_SWAPENTS) +
|
||||||
get_mm_counter(p->mm, MM_SWAPENTS);
|
atomic_long_read(&p->mm->nr_ptes) + mm_nr_pmds(p->mm);
|
||||||
task_unlock(p);
|
task_unlock(p);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -351,7 +351,7 @@ static void dump_tasks(struct mem_cgroup *memcg, const nodemask_t *nodemask)
|
|||||||
struct task_struct *p;
|
struct task_struct *p;
|
||||||
struct task_struct *task;
|
struct task_struct *task;
|
||||||
|
|
||||||
pr_info("[ pid ] uid tgid total_vm rss nr_ptes swapents oom_score_adj name\n");
|
pr_info("[ pid ] uid tgid total_vm rss nr_ptes nr_pmds swapents oom_score_adj name\n");
|
||||||
rcu_read_lock();
|
rcu_read_lock();
|
||||||
for_each_process(p) {
|
for_each_process(p) {
|
||||||
if (oom_unkillable_task(p, memcg, nodemask))
|
if (oom_unkillable_task(p, memcg, nodemask))
|
||||||
@ -367,10 +367,11 @@ static void dump_tasks(struct mem_cgroup *memcg, const nodemask_t *nodemask)
|
|||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
pr_info("[%5d] %5d %5d %8lu %8lu %7ld %8lu %5hd %s\n",
|
pr_info("[%5d] %5d %5d %8lu %8lu %7ld %7ld %8lu %5hd %s\n",
|
||||||
task->pid, from_kuid(&init_user_ns, task_uid(task)),
|
task->pid, from_kuid(&init_user_ns, task_uid(task)),
|
||||||
task->tgid, task->mm->total_vm, get_mm_rss(task->mm),
|
task->tgid, task->mm->total_vm, get_mm_rss(task->mm),
|
||||||
atomic_long_read(&task->mm->nr_ptes),
|
atomic_long_read(&task->mm->nr_ptes),
|
||||||
|
mm_nr_pmds(task->mm),
|
||||||
get_mm_counter(task->mm, MM_SWAPENTS),
|
get_mm_counter(task->mm, MM_SWAPENTS),
|
||||||
task->signal->oom_score_adj, task->comm);
|
task->signal->oom_score_adj, task->comm);
|
||||||
task_unlock(task);
|
task_unlock(task);
|
||||||
|
Loading…
Reference in New Issue
Block a user