243 lines
6.4 KiB
Python
Executable File
243 lines
6.4 KiB
Python
Executable File
#! /usr/bin/python3
|
|
# @lint-avoid-python-3-compatibility-imports
|
|
#
|
|
# slabratetop Summarize kmem_cache_alloc() calls.
|
|
# For Linux, uses BCC, eBPF.
|
|
#
|
|
# USAGE: slabratetop [-h] [-C] [-r MAXROWS] [interval] [count]
|
|
#
|
|
# This uses in-kernel BPF maps to store cache summaries for efficiency.
|
|
#
|
|
# SEE ALSO: slabtop(1), which shows the cache volumes.
|
|
#
|
|
# Copyright 2016 Netflix, Inc.
|
|
# Licensed under the Apache License, Version 2.0 (the "License")
|
|
#
|
|
# 15-Oct-2016 Brendan Gregg Created this.
|
|
# 23-Jan-2023 Rong Tao Introduce kernel internal data structure and
|
|
# functions to temporarily solve problem for
|
|
# >=5.16(TODO: fix this workaround)
|
|
|
|
from __future__ import print_function
|
|
from bcc import BPF
|
|
from bcc.utils import printb
|
|
from time import sleep, strftime
|
|
import argparse
|
|
from subprocess import call
|
|
|
|
# arguments
|
|
examples = """examples:
|
|
./slabratetop # kmem_cache_alloc() top, 1 second refresh
|
|
./slabratetop -C # don't clear the screen
|
|
./slabratetop 5 # 5 second summaries
|
|
./slabratetop 5 10 # 5 second summaries, 10 times only
|
|
"""
|
|
parser = argparse.ArgumentParser(
|
|
description="Kernel SLAB/SLUB memory cache allocation rate top",
|
|
formatter_class=argparse.RawDescriptionHelpFormatter,
|
|
epilog=examples)
|
|
parser.add_argument("-C", "--noclear", action="store_true",
|
|
help="don't clear the screen")
|
|
parser.add_argument("-r", "--maxrows", default=20,
|
|
help="maximum rows to print, default 20")
|
|
parser.add_argument("interval", nargs="?", default=1,
|
|
help="output interval, in seconds")
|
|
parser.add_argument("count", nargs="?", default=99999999,
|
|
help="number of outputs")
|
|
parser.add_argument("--ebpf", action="store_true",
|
|
help=argparse.SUPPRESS)
|
|
args = parser.parse_args()
|
|
interval = int(args.interval)
|
|
countdown = int(args.count)
|
|
maxrows = int(args.maxrows)
|
|
clear = not int(args.noclear)
|
|
debug = 0
|
|
|
|
# linux stats
|
|
loadavg = "/proc/loadavg"
|
|
|
|
# define BPF program
|
|
bpf_text = """
|
|
#include <uapi/linux/ptrace.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/kasan.h>
|
|
|
|
// memcg_cache_params is a part of kmem_cache, but is not publicly exposed in
|
|
// kernel versions 5.4 to 5.8. Define an empty struct for it here to allow the
|
|
// bpf program to compile. It has been completely removed in kernel version
|
|
// 5.9, but it does not hurt to have it here for versions 5.4 to 5.8.
|
|
struct memcg_cache_params {};
|
|
|
|
// introduce kernel interval slab structure and slab_address() function, solved
|
|
// 'undefined' error for >=5.16. TODO: we should fix this workaround if BCC
|
|
// framework support BTF/CO-RE.
|
|
struct slab {
|
|
unsigned long __page_flags;
|
|
|
|
#if defined(CONFIG_SLAB)
|
|
|
|
struct kmem_cache *slab_cache;
|
|
union {
|
|
struct {
|
|
struct list_head slab_list;
|
|
void *freelist; /* array of free object indexes */
|
|
void *s_mem; /* first object */
|
|
};
|
|
struct rcu_head rcu_head;
|
|
};
|
|
unsigned int active;
|
|
|
|
#elif defined(CONFIG_SLUB)
|
|
|
|
struct kmem_cache *slab_cache;
|
|
union {
|
|
struct {
|
|
union {
|
|
struct list_head slab_list;
|
|
#ifdef CONFIG_SLUB_CPU_PARTIAL
|
|
struct {
|
|
struct slab *next;
|
|
int slabs; /* Nr of slabs left */
|
|
};
|
|
#endif
|
|
};
|
|
/* Double-word boundary */
|
|
void *freelist; /* first free object */
|
|
union {
|
|
unsigned long counters;
|
|
struct {
|
|
unsigned inuse:16;
|
|
unsigned objects:15;
|
|
unsigned frozen:1;
|
|
};
|
|
};
|
|
};
|
|
struct rcu_head rcu_head;
|
|
};
|
|
unsigned int __unused;
|
|
|
|
#elif defined(CONFIG_SLOB)
|
|
|
|
struct list_head slab_list;
|
|
void *__unused_1;
|
|
void *freelist; /* first free block */
|
|
long units;
|
|
unsigned int __unused_2;
|
|
|
|
#else
|
|
#error "Unexpected slab allocator configured"
|
|
#endif
|
|
|
|
atomic_t __page_refcount;
|
|
#ifdef CONFIG_MEMCG
|
|
unsigned long memcg_data;
|
|
#endif
|
|
};
|
|
|
|
// slab_address() will not be used, and NULL will be returned directly, which
|
|
// can avoid adaptation of different kernel versions
|
|
static inline void *slab_address(const struct slab *slab)
|
|
{
|
|
return NULL;
|
|
}
|
|
|
|
#ifdef CONFIG_64BIT
|
|
typedef __uint128_t freelist_full_t;
|
|
#else
|
|
typedef u64 freelist_full_t;
|
|
#endif
|
|
|
|
typedef union {
|
|
struct {
|
|
void *freelist;
|
|
unsigned long counter;
|
|
};
|
|
freelist_full_t full;
|
|
} freelist_aba_t;
|
|
|
|
#ifdef CONFIG_SLUB
|
|
#include <linux/slub_def.h>
|
|
#else
|
|
#include <linux/slab_def.h>
|
|
#endif
|
|
|
|
#define CACHE_NAME_SIZE 32
|
|
|
|
// the key for the output summary
|
|
struct info_t {
|
|
char name[CACHE_NAME_SIZE];
|
|
};
|
|
|
|
// the value of the output summary
|
|
struct val_t {
|
|
u64 count;
|
|
u64 size;
|
|
};
|
|
|
|
BPF_HASH(counts, struct info_t, struct val_t);
|
|
|
|
int kprobe__kmem_cache_alloc(struct pt_regs *ctx, struct kmem_cache *cachep)
|
|
{
|
|
struct info_t info = {};
|
|
const char *name = cachep->name;
|
|
bpf_probe_read_kernel(&info.name, sizeof(info.name), name);
|
|
|
|
struct val_t *valp, zero = {};
|
|
valp = counts.lookup_or_try_init(&info, &zero);
|
|
if (valp) {
|
|
valp->count++;
|
|
valp->size += cachep->size;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
"""
|
|
if debug or args.ebpf:
|
|
print(bpf_text)
|
|
if args.ebpf:
|
|
exit()
|
|
|
|
# initialize BPF
|
|
b = BPF(text=bpf_text)
|
|
# check whether hash table batch ops is supported
|
|
htab_batch_ops = True if BPF.kernel_struct_has_field(b'bpf_map_ops',
|
|
b'map_lookup_and_delete_batch') == 1 else False
|
|
|
|
print('Tracing... Output every %d secs. Hit Ctrl-C to end' % interval)
|
|
|
|
# output
|
|
exiting = 0
|
|
while 1:
|
|
try:
|
|
sleep(interval)
|
|
except KeyboardInterrupt:
|
|
exiting = 1
|
|
|
|
# header
|
|
if clear:
|
|
call("clear")
|
|
else:
|
|
print()
|
|
with open(loadavg) as stats:
|
|
print("%-8s loadavg: %s" % (strftime("%H:%M:%S"), stats.read()))
|
|
print("%-32s %6s %10s" % ("CACHE", "ALLOCS", "BYTES"))
|
|
|
|
# by-TID output
|
|
counts = b.get_table("counts")
|
|
line = 0
|
|
for k, v in reversed(sorted(counts.items_lookup_and_delete_batch()
|
|
if htab_batch_ops else counts.items(),
|
|
key=lambda counts: counts[1].size)):
|
|
printb(b"%-32s %6d %10d" % (k.name, v.count, v.size))
|
|
|
|
line += 1
|
|
if line >= maxrows:
|
|
break
|
|
if not htab_batch_ops:
|
|
counts.clear()
|
|
|
|
countdown -= 1
|
|
if exiting or countdown == 0:
|
|
print("Detaching...")
|
|
exit()
|