Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v6.16-rc5 208 lines 4.7 kB view raw
1// SPDX-License-Identifier: GPL-2.0 2#ifndef PERF_LOCK_CONTENTION_H 3#define PERF_LOCK_CONTENTION_H 4 5#include <linux/list.h> 6#include <linux/rbtree.h> 7 8struct lock_filter { 9 int nr_types; 10 int nr_addrs; 11 int nr_syms; 12 int nr_cgrps; 13 int nr_slabs; 14 unsigned int *types; 15 unsigned long *addrs; 16 char **syms; 17 u64 *cgrps; 18 char **slabs; 19}; 20 21struct lock_delay { 22 char *sym; 23 unsigned long addr; 24 unsigned long time; 25}; 26 27struct lock_stat { 28 struct hlist_node hash_entry; 29 struct rb_node rb; /* used for sorting */ 30 31 u64 addr; /* address of lockdep_map, used as ID */ 32 char *name; /* for strcpy(), we cannot use const */ 33 u64 *callstack; 34 35 unsigned int nr_acquire; 36 unsigned int nr_acquired; 37 unsigned int nr_contended; 38 unsigned int nr_release; 39 40 union { 41 unsigned int nr_readlock; 42 unsigned int flags; 43 }; 44 unsigned int nr_trylock; 45 46 /* these times are in nano sec. */ 47 u64 avg_wait_time; 48 u64 wait_time_total; 49 u64 wait_time_min; 50 u64 wait_time_max; 51 52 int broken; /* flag of blacklist */ 53 int combined; 54}; 55 56/* 57 * States of lock_seq_stat 58 * 59 * UNINITIALIZED is required for detecting first event of acquire. 60 * As the nature of lock events, there is no guarantee 61 * that the first event for the locks are acquire, 62 * it can be acquired, contended or release. 63 */ 64#define SEQ_STATE_UNINITIALIZED 0 /* initial state */ 65#define SEQ_STATE_RELEASED 1 66#define SEQ_STATE_ACQUIRING 2 67#define SEQ_STATE_ACQUIRED 3 68#define SEQ_STATE_READ_ACQUIRED 4 69#define SEQ_STATE_CONTENDED 5 70 71/* 72 * MAX_LOCK_DEPTH 73 * Imported from include/linux/sched.h. 74 * Should this be synchronized? 75 */ 76#define MAX_LOCK_DEPTH 48 77 78/* based on kernel/lockdep.c */ 79#define LOCKHASH_BITS 12 80#define LOCKHASH_SIZE (1UL << LOCKHASH_BITS) 81 82extern struct hlist_head *lockhash_table; 83 84/* 85 * struct lock_seq_stat: 86 * Place to put on state of one lock sequence 87 * 1) acquire -> acquired -> release 88 * 2) acquire -> contended -> acquired -> release 89 * 3) acquire (with read or try) -> release 90 * 4) Are there other patterns? 91 */ 92struct lock_seq_stat { 93 struct list_head list; 94 int state; 95 u64 prev_event_time; 96 u64 addr; 97 98 int read_count; 99}; 100 101struct thread_stat { 102 struct rb_node rb; 103 104 u32 tid; 105 struct list_head seq_list; 106}; 107 108/* 109 * CONTENTION_STACK_DEPTH 110 * Number of stack trace entries to find callers 111 */ 112#define CONTENTION_STACK_DEPTH 8 113 114/* 115 * CONTENTION_STACK_SKIP 116 * Number of stack trace entries to skip when finding callers. 117 * The first few entries belong to the locking implementation itself. 118 */ 119#define CONTENTION_STACK_SKIP 4 120 121/* 122 * flags for lock:contention_begin 123 * Imported from include/trace/events/lock.h. 124 */ 125#define LCB_F_SPIN (1U << 0) 126#define LCB_F_READ (1U << 1) 127#define LCB_F_WRITE (1U << 2) 128#define LCB_F_RT (1U << 3) 129#define LCB_F_PERCPU (1U << 4) 130#define LCB_F_MUTEX (1U << 5) 131 132struct evlist; 133struct machine; 134struct target; 135 136struct lock_contention_fails { 137 int task; 138 int stack; 139 int time; 140 int data; 141}; 142 143struct lock_contention { 144 struct evlist *evlist; 145 struct target *target; 146 struct machine *machine; 147 struct hlist_head *result; 148 struct lock_filter *filters; 149 struct lock_delay *delays; 150 struct lock_contention_fails fails; 151 struct rb_root cgroups; 152 void *btf; 153 unsigned long map_nr_entries; 154 int max_stack; 155 int stack_skip; 156 int aggr_mode; 157 int owner; 158 int nr_filtered; 159 int nr_delays; 160 bool save_callstack; 161}; 162 163struct option; 164int parse_call_stack(const struct option *opt, const char *str, int unset); 165bool needs_callstack(void); 166 167struct lock_stat *lock_stat_find(u64 addr); 168struct lock_stat *lock_stat_findnew(u64 addr, const char *name, int flags); 169 170bool match_callstack_filter(struct machine *machine, u64 *callstack, int max_stack_depth); 171 172 173#ifdef HAVE_BPF_SKEL 174int lock_contention_prepare(struct lock_contention *con); 175int lock_contention_start(void); 176int lock_contention_stop(void); 177int lock_contention_read(struct lock_contention *con); 178int lock_contention_finish(struct lock_contention *con); 179 180struct lock_stat *pop_owner_stack_trace(struct lock_contention *con); 181 182#else /* !HAVE_BPF_SKEL */ 183 184static inline int lock_contention_prepare(struct lock_contention *con __maybe_unused) 185{ 186 return 0; 187} 188 189static inline int lock_contention_start(void) { return 0; } 190static inline int lock_contention_stop(void) { return 0; } 191static inline int lock_contention_finish(struct lock_contention *con __maybe_unused) 192{ 193 return 0; 194} 195 196static inline int lock_contention_read(struct lock_contention *con __maybe_unused) 197{ 198 return 0; 199} 200 201static inline struct lock_stat *pop_owner_stack_trace(struct lock_contention *con __maybe_unused) 202{ 203 return NULL; 204} 205 206#endif /* HAVE_BPF_SKEL */ 207 208#endif /* PERF_LOCK_CONTENTION_H */