mirror of
https://github.com/torvalds/linux.git
synced 2024-11-10 14:11:52 +00:00
dee3d0bef2
Reduce the usage of PageFlag tests and reduce the number of compound_head() calls. For multi-page folios, we'll now show all pages as having the flags that apply to them, e.g. if it's dirty, all pages will have the dirty flag set instead of just the head page. The mapped flag is still per page, as is the hwpoison flag. [willy@infradead.org: fix up some bits vs masks] Link: https://lkml.kernel.org/r/20240403173112.1450721-1-willy@infradead.org [willy@infradead.org: fix warnings] Link: https://lkml.kernel.org/r/ZhBPtCYfSuFuUMEz@casper.infradead.org Link: https://lkml.kernel.org/r/20240326171045.410737-11-willy@infradead.org Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org> Reviewed-by: Svetly Todorov <svetly.todorov@memverge.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
226 lines
6.4 KiB
Python
226 lines
6.4 KiB
Python
#!/usr/bin/env drgn
|
|
#
|
|
# Copyright (C) 2020 Roman Gushchin <guro@fb.com>
|
|
# Copyright (C) 2020 Facebook
|
|
|
|
from os import stat
|
|
import argparse
|
|
import sys
|
|
|
|
from drgn.helpers.linux import list_for_each_entry, list_empty
|
|
from drgn.helpers.linux import for_each_page
|
|
from drgn.helpers.linux.cpumask import for_each_online_cpu
|
|
from drgn.helpers.linux.percpu import per_cpu_ptr
|
|
from drgn import container_of, FaultError, Object, cast
|
|
|
|
|
|
DESC = """
|
|
This is a drgn script to provide slab statistics for memory cgroups.
|
|
It supports cgroup v2 and v1 and can emulate memory.kmem.slabinfo
|
|
interface of cgroup v1.
|
|
For drgn, visit https://github.com/osandov/drgn.
|
|
"""
|
|
|
|
|
|
MEMCGS = {}
|
|
|
|
OO_SHIFT = 16
|
|
OO_MASK = ((1 << OO_SHIFT) - 1)
|
|
|
|
|
|
def err(s):
|
|
print('slabinfo.py: error: %s' % s, file=sys.stderr, flush=True)
|
|
sys.exit(1)
|
|
|
|
|
|
def find_memcg_ids(css=prog['root_mem_cgroup'].css, prefix=''):
|
|
if not list_empty(css.children.address_of_()):
|
|
for css in list_for_each_entry('struct cgroup_subsys_state',
|
|
css.children.address_of_(),
|
|
'sibling'):
|
|
name = prefix + '/' + css.cgroup.kn.name.string_().decode('utf-8')
|
|
memcg = container_of(css, 'struct mem_cgroup', 'css')
|
|
MEMCGS[css.cgroup.kn.id.value_()] = memcg
|
|
find_memcg_ids(css, name)
|
|
|
|
|
|
def is_root_cache(s):
|
|
try:
|
|
return False if s.memcg_params.root_cache else True
|
|
except AttributeError:
|
|
return True
|
|
|
|
|
|
def cache_name(s):
|
|
if is_root_cache(s):
|
|
return s.name.string_().decode('utf-8')
|
|
else:
|
|
return s.memcg_params.root_cache.name.string_().decode('utf-8')
|
|
|
|
|
|
# SLUB
|
|
|
|
def oo_order(s):
|
|
return s.oo.x >> OO_SHIFT
|
|
|
|
|
|
def oo_objects(s):
|
|
return s.oo.x & OO_MASK
|
|
|
|
|
|
def count_partial(n, fn):
|
|
nr_objs = 0
|
|
for slab in list_for_each_entry('struct slab', n.partial.address_of_(),
|
|
'slab_list'):
|
|
nr_objs += fn(slab)
|
|
return nr_objs
|
|
|
|
|
|
def count_free(slab):
|
|
return slab.objects - slab.inuse
|
|
|
|
|
|
def slub_get_slabinfo(s, cfg):
|
|
nr_slabs = 0
|
|
nr_objs = 0
|
|
nr_free = 0
|
|
|
|
for node in range(cfg['nr_nodes']):
|
|
n = s.node[node]
|
|
nr_slabs += n.nr_slabs.counter.value_()
|
|
nr_objs += n.total_objects.counter.value_()
|
|
nr_free += count_partial(n, count_free)
|
|
|
|
return {'active_objs': nr_objs - nr_free,
|
|
'num_objs': nr_objs,
|
|
'active_slabs': nr_slabs,
|
|
'num_slabs': nr_slabs,
|
|
'objects_per_slab': oo_objects(s),
|
|
'cache_order': oo_order(s),
|
|
'limit': 0,
|
|
'batchcount': 0,
|
|
'shared': 0,
|
|
'shared_avail': 0}
|
|
|
|
|
|
def cache_show(s, cfg, objs):
|
|
if cfg['allocator'] == 'SLUB':
|
|
sinfo = slub_get_slabinfo(s, cfg)
|
|
else:
|
|
err('SLAB isn\'t supported yet')
|
|
|
|
if cfg['shared_slab_pages']:
|
|
sinfo['active_objs'] = objs
|
|
sinfo['num_objs'] = objs
|
|
|
|
print('%-17s %6lu %6lu %6u %4u %4d'
|
|
' : tunables %4u %4u %4u'
|
|
' : slabdata %6lu %6lu %6lu' % (
|
|
cache_name(s), sinfo['active_objs'], sinfo['num_objs'],
|
|
s.size, sinfo['objects_per_slab'], 1 << sinfo['cache_order'],
|
|
sinfo['limit'], sinfo['batchcount'], sinfo['shared'],
|
|
sinfo['active_slabs'], sinfo['num_slabs'],
|
|
sinfo['shared_avail']))
|
|
|
|
|
|
def detect_kernel_config():
|
|
cfg = {}
|
|
|
|
cfg['nr_nodes'] = prog['nr_online_nodes'].value_()
|
|
|
|
if prog.type('struct kmem_cache').members[1].name == 'flags':
|
|
cfg['allocator'] = 'SLUB'
|
|
elif prog.type('struct kmem_cache').members[1].name == 'batchcount':
|
|
cfg['allocator'] = 'SLAB'
|
|
else:
|
|
err('Can\'t determine the slab allocator')
|
|
|
|
cfg['shared_slab_pages'] = False
|
|
try:
|
|
if prog.type('struct obj_cgroup'):
|
|
cfg['shared_slab_pages'] = True
|
|
except:
|
|
pass
|
|
|
|
return cfg
|
|
|
|
|
|
def for_each_slab(prog):
|
|
PGSlab = ~prog.constant('PG_slab')
|
|
|
|
for page in for_each_page(prog):
|
|
try:
|
|
if page.page_type.value_() == PGSlab:
|
|
yield cast('struct slab *', page)
|
|
except FaultError:
|
|
pass
|
|
|
|
|
|
def main():
|
|
parser = argparse.ArgumentParser(description=DESC,
|
|
formatter_class=
|
|
argparse.RawTextHelpFormatter)
|
|
parser.add_argument('cgroup', metavar='CGROUP',
|
|
help='Target memory cgroup')
|
|
args = parser.parse_args()
|
|
|
|
try:
|
|
cgroup_id = stat(args.cgroup).st_ino
|
|
find_memcg_ids()
|
|
memcg = MEMCGS[cgroup_id]
|
|
except KeyError:
|
|
err('Can\'t find the memory cgroup')
|
|
|
|
cfg = detect_kernel_config()
|
|
|
|
print('# name <active_objs> <num_objs> <objsize> <objperslab> <pagesperslab>'
|
|
' : tunables <limit> <batchcount> <sharedfactor>'
|
|
' : slabdata <active_slabs> <num_slabs> <sharedavail>')
|
|
|
|
if cfg['shared_slab_pages']:
|
|
obj_cgroups = set()
|
|
stats = {}
|
|
caches = {}
|
|
|
|
# find memcg pointers belonging to the specified cgroup
|
|
obj_cgroups.add(memcg.objcg.value_())
|
|
for ptr in list_for_each_entry('struct obj_cgroup',
|
|
memcg.objcg_list.address_of_(),
|
|
'list'):
|
|
obj_cgroups.add(ptr.value_())
|
|
|
|
# look over all slab folios and look for objects belonging
|
|
# to the given memory cgroup
|
|
for slab in for_each_slab(prog):
|
|
objcg_vec_raw = slab.memcg_data.value_()
|
|
if objcg_vec_raw == 0:
|
|
continue
|
|
cache = slab.slab_cache
|
|
if not cache:
|
|
continue
|
|
addr = cache.value_()
|
|
caches[addr] = cache
|
|
# clear the lowest bit to get the true obj_cgroups
|
|
objcg_vec = Object(prog, 'struct obj_cgroup **',
|
|
value=objcg_vec_raw & ~1)
|
|
|
|
if addr not in stats:
|
|
stats[addr] = 0
|
|
|
|
for i in range(oo_objects(cache)):
|
|
if objcg_vec[i].value_() in obj_cgroups:
|
|
stats[addr] += 1
|
|
|
|
for addr in caches:
|
|
if stats[addr] > 0:
|
|
cache_show(caches[addr], cfg, stats[addr])
|
|
|
|
else:
|
|
for s in list_for_each_entry('struct kmem_cache',
|
|
memcg.kmem_caches.address_of_(),
|
|
'memcg_params.kmem_caches_node'):
|
|
cache_show(s, cfg, None)
|
|
|
|
|
|
main()
|