Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1// SPDX-License-Identifier: GPL-2.0
2#ifndef PERF_LOCK_CONTENTION_H
3#define PERF_LOCK_CONTENTION_H
4
5#include <linux/list.h>
6#include <linux/rbtree.h>
7
8struct lock_filter {
9 int nr_types;
10 int nr_addrs;
11 int nr_syms;
12 int nr_cgrps;
13 int nr_slabs;
14 unsigned int *types;
15 unsigned long *addrs;
16 char **syms;
17 u64 *cgrps;
18 char **slabs;
19};
20
21struct lock_delay {
22 char *sym;
23 unsigned long addr;
24 unsigned long time;
25};
26
27struct lock_stat {
28 struct hlist_node hash_entry;
29 struct rb_node rb; /* used for sorting */
30
31 u64 addr; /* address of lockdep_map, used as ID */
32 char *name; /* for strcpy(), we cannot use const */
33 u64 *callstack;
34
35 unsigned int nr_acquire;
36 unsigned int nr_acquired;
37 unsigned int nr_contended;
38 unsigned int nr_release;
39
40 union {
41 unsigned int nr_readlock;
42 unsigned int flags;
43 };
44 unsigned int nr_trylock;
45
46 /* these times are in nano sec. */
47 u64 avg_wait_time;
48 u64 wait_time_total;
49 u64 wait_time_min;
50 u64 wait_time_max;
51
52 int broken; /* flag of blacklist */
53 int combined;
54};
55
56/*
57 * States of lock_seq_stat
58 *
59 * UNINITIALIZED is required for detecting first event of acquire.
60 * As the nature of lock events, there is no guarantee
61 * that the first event for the locks are acquire,
62 * it can be acquired, contended or release.
63 */
64#define SEQ_STATE_UNINITIALIZED 0 /* initial state */
65#define SEQ_STATE_RELEASED 1
66#define SEQ_STATE_ACQUIRING 2
67#define SEQ_STATE_ACQUIRED 3
68#define SEQ_STATE_READ_ACQUIRED 4
69#define SEQ_STATE_CONTENDED 5
70
71/*
72 * MAX_LOCK_DEPTH
73 * Imported from include/linux/sched.h.
74 * Should this be synchronized?
75 */
76#define MAX_LOCK_DEPTH 48
77
78/* based on kernel/lockdep.c */
79#define LOCKHASH_BITS 12
80#define LOCKHASH_SIZE (1UL << LOCKHASH_BITS)
81
82extern struct hlist_head *lockhash_table;
83
84/*
85 * struct lock_seq_stat:
86 * Place to put on state of one lock sequence
87 * 1) acquire -> acquired -> release
88 * 2) acquire -> contended -> acquired -> release
89 * 3) acquire (with read or try) -> release
90 * 4) Are there other patterns?
91 */
92struct lock_seq_stat {
93 struct list_head list;
94 int state;
95 u64 prev_event_time;
96 u64 addr;
97
98 int read_count;
99};
100
101struct thread_stat {
102 struct rb_node rb;
103
104 u32 tid;
105 struct list_head seq_list;
106};
107
108/*
109 * CONTENTION_STACK_DEPTH
110 * Number of stack trace entries to find callers
111 */
112#define CONTENTION_STACK_DEPTH 8
113
114/*
115 * CONTENTION_STACK_SKIP
116 * Number of stack trace entries to skip when finding callers.
117 * The first few entries belong to the locking implementation itself.
118 */
119#define CONTENTION_STACK_SKIP 4
120
121/*
122 * flags for lock:contention_begin
123 * Imported from include/trace/events/lock.h.
124 */
125#define LCB_F_SPIN (1U << 0)
126#define LCB_F_READ (1U << 1)
127#define LCB_F_WRITE (1U << 2)
128#define LCB_F_RT (1U << 3)
129#define LCB_F_PERCPU (1U << 4)
130#define LCB_F_MUTEX (1U << 5)
131
132struct evlist;
133struct machine;
134struct target;
135
136struct lock_contention_fails {
137 int task;
138 int stack;
139 int time;
140 int data;
141};
142
143struct lock_contention {
144 struct evlist *evlist;
145 struct target *target;
146 struct machine *machine;
147 struct hlist_head *result;
148 struct lock_filter *filters;
149 struct lock_delay *delays;
150 struct lock_contention_fails fails;
151 struct rb_root cgroups;
152 void *btf;
153 unsigned long map_nr_entries;
154 int max_stack;
155 int stack_skip;
156 int aggr_mode;
157 int owner;
158 int nr_filtered;
159 int nr_delays;
160 bool save_callstack;
161};
162
163struct option;
164int parse_call_stack(const struct option *opt, const char *str, int unset);
165bool needs_callstack(void);
166
167struct lock_stat *lock_stat_find(u64 addr);
168struct lock_stat *lock_stat_findnew(u64 addr, const char *name, int flags);
169
170bool match_callstack_filter(struct machine *machine, u64 *callstack, int max_stack_depth);
171
172
173#ifdef HAVE_BPF_SKEL
174int lock_contention_prepare(struct lock_contention *con);
175int lock_contention_start(void);
176int lock_contention_stop(void);
177int lock_contention_read(struct lock_contention *con);
178int lock_contention_finish(struct lock_contention *con);
179
180struct lock_stat *pop_owner_stack_trace(struct lock_contention *con);
181
182#else /* !HAVE_BPF_SKEL */
183
184static inline int lock_contention_prepare(struct lock_contention *con __maybe_unused)
185{
186 return 0;
187}
188
189static inline int lock_contention_start(void) { return 0; }
190static inline int lock_contention_stop(void) { return 0; }
191static inline int lock_contention_finish(struct lock_contention *con __maybe_unused)
192{
193 return 0;
194}
195
196static inline int lock_contention_read(struct lock_contention *con __maybe_unused)
197{
198 return 0;
199}
200
201static inline struct lock_stat *pop_owner_stack_trace(struct lock_contention *con __maybe_unused)
202{
203 return NULL;
204}
205
206#endif /* HAVE_BPF_SKEL */
207
208#endif /* PERF_LOCK_CONTENTION_H */