%PDF- %PDF-
Direktori : /usr/sbin/ |
Current File : //usr/sbin/slabratetop-bpfcc |
#! /usr/bin/python3 # @lint-avoid-python-3-compatibility-imports # # slabratetop Summarize kmem_cache_alloc() calls. # For Linux, uses BCC, eBPF. # # USAGE: slabratetop [-h] [-C] [-r MAXROWS] [interval] [count] # # This uses in-kernel BPF maps to store cache summaries for efficiency. # # SEE ALSO: slabtop(1), which shows the cache volumes. # # Copyright 2016 Netflix, Inc. # Licensed under the Apache License, Version 2.0 (the "License") # # 15-Oct-2016 Brendan Gregg Created this. # 23-Jan-2023 Rong Tao Introduce kernel internal data structure and # functions to temporarily solve problem for # >=5.16(TODO: fix this workaround) from __future__ import print_function from bcc import BPF from bcc.utils import printb from time import sleep, strftime import argparse from subprocess import call # arguments examples = """examples: ./slabratetop # kmem_cache_alloc() top, 1 second refresh ./slabratetop -C # don't clear the screen ./slabratetop 5 # 5 second summaries ./slabratetop 5 10 # 5 second summaries, 10 times only """ parser = argparse.ArgumentParser( description="Kernel SLAB/SLUB memory cache allocation rate top", formatter_class=argparse.RawDescriptionHelpFormatter, epilog=examples) parser.add_argument("-C", "--noclear", action="store_true", help="don't clear the screen") parser.add_argument("-r", "--maxrows", default=20, help="maximum rows to print, default 20") parser.add_argument("interval", nargs="?", default=1, help="output interval, in seconds") parser.add_argument("count", nargs="?", default=99999999, help="number of outputs") parser.add_argument("--ebpf", action="store_true", help=argparse.SUPPRESS) args = parser.parse_args() interval = int(args.interval) countdown = int(args.count) maxrows = int(args.maxrows) clear = not int(args.noclear) debug = 0 # linux stats loadavg = "/proc/loadavg" # define BPF program bpf_text = """ #include <uapi/linux/ptrace.h> #include <linux/mm.h> #include <linux/kasan.h> // memcg_cache_params is a part of kmem_cache, but is not publicly exposed in // kernel versions 5.4 to 5.8. Define an empty struct for it here to allow the // bpf program to compile. It has been completely removed in kernel version // 5.9, but it does not hurt to have it here for versions 5.4 to 5.8. struct memcg_cache_params {}; // introduce kernel interval slab structure and slab_address() function, solved // 'undefined' error for >=5.16. TODO: we should fix this workaround if BCC // framework support BTF/CO-RE. struct slab { unsigned long __page_flags; #if defined(CONFIG_SLAB) struct kmem_cache *slab_cache; union { struct { struct list_head slab_list; void *freelist; /* array of free object indexes */ void *s_mem; /* first object */ }; struct rcu_head rcu_head; }; unsigned int active; #elif defined(CONFIG_SLUB) struct kmem_cache *slab_cache; union { struct { union { struct list_head slab_list; #ifdef CONFIG_SLUB_CPU_PARTIAL struct { struct slab *next; int slabs; /* Nr of slabs left */ }; #endif }; /* Double-word boundary */ void *freelist; /* first free object */ union { unsigned long counters; struct { unsigned inuse:16; unsigned objects:15; unsigned frozen:1; }; }; }; struct rcu_head rcu_head; }; unsigned int __unused; #elif defined(CONFIG_SLOB) struct list_head slab_list; void *__unused_1; void *freelist; /* first free block */ long units; unsigned int __unused_2; #else #error "Unexpected slab allocator configured" #endif atomic_t __page_refcount; #ifdef CONFIG_MEMCG unsigned long memcg_data; #endif }; // slab_address() will not be used, and NULL will be returned directly, which // can avoid adaptation of different kernel versions static inline void *slab_address(const struct slab *slab) { return NULL; } #ifdef CONFIG_64BIT typedef __uint128_t freelist_full_t; #else typedef u64 freelist_full_t; #endif typedef union { struct { void *freelist; unsigned long counter; }; freelist_full_t full; } freelist_aba_t; #ifdef CONFIG_SLUB #include <linux/slub_def.h> #else #include <linux/slab_def.h> #endif #define CACHE_NAME_SIZE 32 // the key for the output summary struct info_t { char name[CACHE_NAME_SIZE]; }; // the value of the output summary struct val_t { u64 count; u64 size; }; BPF_HASH(counts, struct info_t, struct val_t); int kprobe__kmem_cache_alloc(struct pt_regs *ctx, struct kmem_cache *cachep) { struct info_t info = {}; const char *name = cachep->name; bpf_probe_read_kernel(&info.name, sizeof(info.name), name); struct val_t *valp, zero = {}; valp = counts.lookup_or_try_init(&info, &zero); if (valp) { valp->count++; valp->size += cachep->size; } return 0; } """ if debug or args.ebpf: print(bpf_text) if args.ebpf: exit() # initialize BPF b = BPF(text=bpf_text) # check whether hash table batch ops is supported htab_batch_ops = True if BPF.kernel_struct_has_field(b'bpf_map_ops', b'map_lookup_and_delete_batch') == 1 else False print('Tracing... Output every %d secs. Hit Ctrl-C to end' % interval) # output exiting = 0 while 1: try: sleep(interval) except KeyboardInterrupt: exiting = 1 # header if clear: call("clear") else: print() with open(loadavg) as stats: print("%-8s loadavg: %s" % (strftime("%H:%M:%S"), stats.read())) print("%-32s %6s %10s" % ("CACHE", "ALLOCS", "BYTES")) # by-TID output counts = b.get_table("counts") line = 0 for k, v in reversed(sorted(counts.items_lookup_and_delete_batch() if htab_batch_ops else counts.items(), key=lambda counts: counts[1].size)): printb(b"%-32s %6d %10d" % (k.name, v.count, v.size)) line += 1 if line >= maxrows: break if not htab_batch_ops: counts.clear() countdown -= 1 if exiting or countdown == 0: print("Detaching...") exit()