%PDF- %PDF-
Direktori : /usr/sbin/ |
Current File : //usr/sbin/dirtop-bpfcc |
#! /usr/bin/python3 # @lint-avoid-python-3-compatibility-imports # # dirtop file reads and writes by directory. # For Linux, uses BCC, eBPF. # # USAGE: dirtop.py -d 'directory1,directory2' [-h] [-C] [-r MAXROWS] [interval] [count] # # This uses in-kernel eBPF maps to store per process summaries for efficiency. # # Copyright 2016 Netflix, Inc. # Licensed under the Apache License, Version 2.0 (the "License") # # 13-Mar-2020 Erwan Velu Created dirtop from filetop # 06-Feb-2016 Brendan Gregg Created filetop. from __future__ import print_function from bcc import BPF from time import sleep, strftime import argparse import os import stat from subprocess import call # arguments examples = """examples: ./dirtop -d '/hdfs/uuid/*/yarn' # directory I/O top, 1 second refresh ./dirtop -d '/hdfs/uuid/*/yarn' -C # don't clear the screen ./dirtop -d '/hdfs/uuid/*/yarn' 5 # 5 second summaries ./dirtop -d '/hdfs/uuid/*/yarn' 5 10 # 5 second summaries, 10 times only ./dirtop -d '/hdfs/uuid/*/yarn,/hdfs/uuid/*/data' # Running dirtop on two set of directories """ parser = argparse.ArgumentParser( description="File reads and writes by process", formatter_class=argparse.RawDescriptionHelpFormatter, epilog=examples) parser.add_argument("-C", "--noclear", action="store_true", help="don't clear the screen") parser.add_argument("-r", "--maxrows", default=20, help="maximum rows to print, default 20") parser.add_argument("-s", "--sort", default="all", choices=["all", "reads", "writes", "rbytes", "wbytes"], help="sort column, default all") parser.add_argument("-p", "--pid", type=int, metavar="PID", dest="tgid", help="trace this PID only") parser.add_argument("interval", nargs="?", default=1, help="output interval, in seconds") parser.add_argument("count", nargs="?", default=99999999, help="number of outputs") parser.add_argument("--ebpf", action="store_true", help=argparse.SUPPRESS) parser.add_argument("-d", "--root-directories", type=str, required=True, dest="rootdirs", help="select the directories to observe, separated by commas") args = parser.parse_args() interval = int(args.interval) countdown = int(args.count) maxrows = int(args.maxrows) clear = not int(args.noclear) debug = 0 # linux stats loadavg = "/proc/loadavg" # define BPF program bpf_text = """ # include <uapi/linux/ptrace.h> # include <linux/blkdev.h> // the key for the output summary struct info_t { unsigned long inode_id; }; // the value of the output summary struct val_t { u64 reads; u64 writes; u64 rbytes; u64 wbytes; }; BPF_HASH(counts, struct info_t, struct val_t); static int do_entry(struct pt_regs *ctx, struct file *file, char __user *buf, size_t count, int is_read) { u32 tgid = bpf_get_current_pid_tgid() >> 32; if (TGID_FILTER) return 0; // The directory inodes we look at u32 dir_ids[INODES_NUMBER] = DIRECTORY_INODES; struct info_t info = {.inode_id = 0}; struct dentry *pde = file->f_path.dentry; for (int i=0; i<50; i++) { // If we don't have any parent, we reached the root if (!pde->d_parent) { break; } pde = pde->d_parent; // Does the files is part of the directory we look for for(int dir_id=0; dir_id<INODES_NUMBER; dir_id++) { if (pde->d_inode->i_ino == dir_ids[dir_id]) { // Yes, let's export the top directory inode info.inode_id = pde->d_inode->i_ino; break; } } } // If we didn't found any, let's abort if (info.inode_id == 0) { return 0; } struct val_t *valp, zero = {}; valp = counts.lookup_or_try_init(&info, &zero); if (valp) { if (is_read) { valp->reads++; valp->rbytes += count; } else { valp->writes++; valp->wbytes += count; } } return 0; } int trace_read_entry(struct pt_regs *ctx, struct file *file, char __user *buf, size_t count) { return do_entry(ctx, file, buf, count, 1); } int trace_write_entry(struct pt_regs *ctx, struct file *file, char __user *buf, size_t count) { return do_entry(ctx, file, buf, count, 0); } """ def get_searched_ids(root_directories): """Export the inode numbers of the selected directories.""" from glob import glob inode_to_path = {} inodes = "{" total_dirs = 0 for root_directory in root_directories.split(','): try: searched_dirs = glob(root_directory, recursive=True) except TypeError: searched_dirs = glob(root_directory) if not searched_dirs: continue for mydir in searched_dirs: total_dirs = total_dirs + 1 # If we pass more than 15 dirs, ebpf program fails if total_dirs > 15: print('15 directories limit reached') break inode_id = os.lstat(mydir)[stat.ST_INO] if inode_id in inode_to_path: if inode_to_path[inode_id] == mydir: print('Skipping {} as already considered'.format(mydir)) else: inodes = "{},{}".format(inodes, inode_id) inode_to_path[inode_id] = mydir print('Considering {} with inode_id {}'.format(mydir, inode_id)) inodes = inodes + '}' if len(inode_to_path) == 0: print('Cannot find any valid directory') exit() return inodes.replace('{,', '{'), inode_to_path if args.tgid: bpf_text = bpf_text.replace('TGID_FILTER', 'tgid != %d' % args.tgid) else: bpf_text = bpf_text.replace('TGID_FILTER', '0') inodes, inodes_to_path = get_searched_ids(args.rootdirs) bpf_text = bpf_text.replace("DIRECTORY_INODES", inodes) bpf_text = bpf_text.replace( "INODES_NUMBER", '{}'.format(len(inodes.split(',')))) if debug or args.ebpf: print(bpf_text) if args.ebpf: exit() # initialize BPF b = BPF(text=bpf_text) b.attach_kprobe(event="vfs_read", fn_name="trace_read_entry") b.attach_kprobe(event="vfs_write", fn_name="trace_write_entry") # check whether hash table batch ops is supported htab_batch_ops = True if BPF.kernel_struct_has_field(b'bpf_map_ops', b'map_lookup_and_delete_batch') == 1 else False DNAME_INLINE_LEN = 32 # linux/dcache.h print('Tracing... Output every %d secs. Hit Ctrl-C to end' % interval) def sort_fn(counts): """Define how to sort the columns""" if args.sort == "all": return (counts[1].rbytes + counts[1].wbytes + counts[1].reads + counts[1].writes) else: return getattr(counts[1], args.sort) # output exiting = 0 while 1: try: sleep(interval) except KeyboardInterrupt: exiting = 1 # header if clear: call("clear") else: print() with open(loadavg) as stats: print("%-8s loadavg: %s" % (strftime("%H:%M:%S"), stats.read())) print("%-6s %-6s %-8s %-8s %s" % ("READS", "WRITES", "R_Kb", "W_Kb", "PATH")) # by-TID output counts = b.get_table("counts") line = 0 reads = {} writes = {} reads_Kb = {} writes_Kb = {} for k, v in reversed(sorted(counts.items_lookup_and_delete_batch() if htab_batch_ops else counts.items(), key=sort_fn)): # If it's the first time we see this inode if k.inode_id not in reads: # let's create a new entry reads[k.inode_id] = v.reads writes[k.inode_id] = v.writes reads_Kb[k.inode_id] = v.rbytes / 1024 writes_Kb[k.inode_id] = v.wbytes / 1024 else: # unless add the current performance metrics # to the previous ones reads[k.inode_id] += v.reads writes[k.inode_id] += v.writes reads_Kb[k.inode_id] += v.rbytes / 1024 writes_Kb[k.inode_id] += v.wbytes / 1024 for node_id in reads: print("%-6d %-6d %-8d %-8d %s" % (reads[node_id], writes[node_id], reads_Kb[node_id], writes_Kb[node_id], inodes_to_path[node_id])) line += 1 if line >= maxrows: break if not htab_batch_ops: counts.clear() countdown -= 1 if exiting or countdown == 0: print("Detaching...") exit()