|  | #!/usr/bin/env drgn | 
|  | # | 
|  | # Copyright (C) 2020 Roman Gushchin <guro@fb.com> | 
|  | # Copyright (C) 2020 Facebook | 
|  |  | 
|  | from os import stat | 
|  | import argparse | 
|  | import sys | 
|  |  | 
|  | from drgn.helpers.linux import list_for_each_entry, list_empty | 
|  | from drgn.helpers.linux import for_each_page | 
|  | from drgn.helpers.linux.cpumask import for_each_online_cpu | 
|  | from drgn.helpers.linux.percpu import per_cpu_ptr | 
|  | from drgn import container_of, FaultError, Object, cast | 
|  |  | 
|  |  | 
|  | DESC = """ | 
|  | This is a drgn script to provide slab statistics for memory cgroups. | 
|  | It supports cgroup v2 and v1 and can emulate memory.kmem.slabinfo | 
|  | interface of cgroup v1. | 
|  | For drgn, visit https://github.com/osandov/drgn. | 
|  | """ | 
|  |  | 
|  |  | 
|  | MEMCGS = {} | 
|  |  | 
|  | OO_SHIFT = 16 | 
|  | OO_MASK = ((1 << OO_SHIFT) - 1) | 
|  |  | 
|  |  | 
|  | def err(s): | 
|  | print('slabinfo.py: error: %s' % s, file=sys.stderr, flush=True) | 
|  | sys.exit(1) | 
|  |  | 
|  |  | 
|  | def find_memcg_ids(css=prog['root_mem_cgroup'].css, prefix=''): | 
|  | if not list_empty(css.children.address_of_()): | 
|  | for css in list_for_each_entry('struct cgroup_subsys_state', | 
|  | css.children.address_of_(), | 
|  | 'sibling'): | 
|  | name = prefix + '/' + css.cgroup.kn.name.string_().decode('utf-8') | 
|  | memcg = container_of(css, 'struct mem_cgroup', 'css') | 
|  | MEMCGS[css.cgroup.kn.id.value_()] = memcg | 
|  | find_memcg_ids(css, name) | 
|  |  | 
|  |  | 
|  | def is_root_cache(s): | 
|  | try: | 
|  | return False if s.memcg_params.root_cache else True | 
|  | except AttributeError: | 
|  | return True | 
|  |  | 
|  |  | 
|  | def cache_name(s): | 
|  | if is_root_cache(s): | 
|  | return s.name.string_().decode('utf-8') | 
|  | else: | 
|  | return s.memcg_params.root_cache.name.string_().decode('utf-8') | 
|  |  | 
|  |  | 
|  | # SLUB | 
|  |  | 
|  | def oo_order(s): | 
|  | return s.oo.x >> OO_SHIFT | 
|  |  | 
|  |  | 
|  | def oo_objects(s): | 
|  | return s.oo.x & OO_MASK | 
|  |  | 
|  |  | 
|  | def count_partial(n, fn): | 
|  | nr_objs = 0 | 
|  | for slab in list_for_each_entry('struct slab', n.partial.address_of_(), | 
|  | 'slab_list'): | 
|  | nr_objs += fn(slab) | 
|  | return nr_objs | 
|  |  | 
|  |  | 
|  | def count_free(slab): | 
|  | return slab.objects - slab.inuse | 
|  |  | 
|  |  | 
|  | def slub_get_slabinfo(s, cfg): | 
|  | nr_slabs = 0 | 
|  | nr_objs = 0 | 
|  | nr_free = 0 | 
|  |  | 
|  | for node in range(cfg['nr_nodes']): | 
|  | n = s.node[node] | 
|  | nr_slabs += n.nr_slabs.counter.value_() | 
|  | nr_objs += n.total_objects.counter.value_() | 
|  | nr_free += count_partial(n, count_free) | 
|  |  | 
|  | return {'active_objs': nr_objs - nr_free, | 
|  | 'num_objs': nr_objs, | 
|  | 'active_slabs': nr_slabs, | 
|  | 'num_slabs': nr_slabs, | 
|  | 'objects_per_slab': oo_objects(s), | 
|  | 'cache_order': oo_order(s), | 
|  | 'limit': 0, | 
|  | 'batchcount': 0, | 
|  | 'shared': 0, | 
|  | 'shared_avail': 0} | 
|  |  | 
|  |  | 
|  | def cache_show(s, cfg, objs): | 
|  | if cfg['allocator'] == 'SLUB': | 
|  | sinfo = slub_get_slabinfo(s, cfg) | 
|  | else: | 
|  | err('SLAB isn\'t supported yet') | 
|  |  | 
|  | if cfg['shared_slab_pages']: | 
|  | sinfo['active_objs'] = objs | 
|  | sinfo['num_objs'] = objs | 
|  |  | 
|  | print('%-17s %6lu %6lu %6u %4u %4d' | 
|  | ' : tunables %4u %4u %4u' | 
|  | ' : slabdata %6lu %6lu %6lu' % ( | 
|  | cache_name(s), sinfo['active_objs'], sinfo['num_objs'], | 
|  | s.size, sinfo['objects_per_slab'], 1 << sinfo['cache_order'], | 
|  | sinfo['limit'], sinfo['batchcount'], sinfo['shared'], | 
|  | sinfo['active_slabs'], sinfo['num_slabs'], | 
|  | sinfo['shared_avail'])) | 
|  |  | 
|  |  | 
|  | def detect_kernel_config(): | 
|  | cfg = {} | 
|  |  | 
|  | cfg['nr_nodes'] = prog['nr_online_nodes'].value_() | 
|  |  | 
|  | if prog.type('struct kmem_cache').members[1].name == 'flags': | 
|  | cfg['allocator'] = 'SLUB' | 
|  | elif prog.type('struct kmem_cache').members[1].name == 'batchcount': | 
|  | cfg['allocator'] = 'SLAB' | 
|  | else: | 
|  | err('Can\'t determine the slab allocator') | 
|  |  | 
|  | cfg['shared_slab_pages'] = False | 
|  | try: | 
|  | if prog.type('struct obj_cgroup'): | 
|  | cfg['shared_slab_pages'] = True | 
|  | except: | 
|  | pass | 
|  |  | 
|  | return cfg | 
|  |  | 
|  |  | 
|  | def for_each_slab(prog): | 
|  | slabtype = prog.constant('PGTY_slab') | 
|  |  | 
|  | for page in for_each_page(prog): | 
|  | try: | 
|  | if (page.page_type.value_() >> 24) == slabtype: | 
|  | yield cast('struct slab *', page) | 
|  | except FaultError: | 
|  | pass | 
|  |  | 
|  |  | 
|  | def main(): | 
|  | parser = argparse.ArgumentParser(description=DESC, | 
|  | formatter_class= | 
|  | argparse.RawTextHelpFormatter) | 
|  | parser.add_argument('cgroup', metavar='CGROUP', | 
|  | help='Target memory cgroup') | 
|  | args = parser.parse_args() | 
|  |  | 
|  | try: | 
|  | cgroup_id = stat(args.cgroup).st_ino | 
|  | find_memcg_ids() | 
|  | memcg = MEMCGS[cgroup_id] | 
|  | except KeyError: | 
|  | err('Can\'t find the memory cgroup') | 
|  |  | 
|  | cfg = detect_kernel_config() | 
|  |  | 
|  | print('# name            <active_objs> <num_objs> <objsize> <objperslab> <pagesperslab>' | 
|  | ' : tunables <limit> <batchcount> <sharedfactor>' | 
|  | ' : slabdata <active_slabs> <num_slabs> <sharedavail>') | 
|  |  | 
|  | if cfg['shared_slab_pages']: | 
|  | obj_cgroups = set() | 
|  | stats = {} | 
|  | caches = {} | 
|  |  | 
|  | # find memcg pointers belonging to the specified cgroup | 
|  | obj_cgroups.add(memcg.objcg.value_()) | 
|  | for ptr in list_for_each_entry('struct obj_cgroup', | 
|  | memcg.objcg_list.address_of_(), | 
|  | 'list'): | 
|  | obj_cgroups.add(ptr.value_()) | 
|  |  | 
|  | # look over all slab folios and look for objects belonging | 
|  | # to the given memory cgroup | 
|  | for slab in for_each_slab(prog): | 
|  | objcg_vec_raw = slab.memcg_data.value_() | 
|  | if objcg_vec_raw == 0: | 
|  | continue | 
|  | cache = slab.slab_cache | 
|  | if not cache: | 
|  | continue | 
|  | addr = cache.value_() | 
|  | caches[addr] = cache | 
|  | # clear the lowest bit to get the true obj_cgroups | 
|  | objcg_vec = Object(prog, 'struct obj_cgroup **', | 
|  | value=objcg_vec_raw & ~1) | 
|  |  | 
|  | if addr not in stats: | 
|  | stats[addr] = 0 | 
|  |  | 
|  | for i in range(oo_objects(cache)): | 
|  | if objcg_vec[i].value_() in obj_cgroups: | 
|  | stats[addr] += 1 | 
|  |  | 
|  | for addr in caches: | 
|  | if stats[addr] > 0: | 
|  | cache_show(caches[addr], cfg, stats[addr]) | 
|  |  | 
|  | else: | 
|  | for s in list_for_each_entry('struct kmem_cache', | 
|  | memcg.kmem_caches.address_of_(), | 
|  | 'memcg_params.kmem_caches_node'): | 
|  | cache_show(s, cfg, None) | 
|  |  | 
|  |  | 
|  | main() |