at v4.13 15 kB view raw
1#ifndef _LINUX_KPROBES_H 2#define _LINUX_KPROBES_H 3/* 4 * Kernel Probes (KProbes) 5 * include/linux/kprobes.h 6 * 7 * This program is free software; you can redistribute it and/or modify 8 * it under the terms of the GNU General Public License as published by 9 * the Free Software Foundation; either version 2 of the License, or 10 * (at your option) any later version. 11 * 12 * This program is distributed in the hope that it will be useful, 13 * but WITHOUT ANY WARRANTY; without even the implied warranty of 14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 15 * GNU General Public License for more details. 16 * 17 * You should have received a copy of the GNU General Public License 18 * along with this program; if not, write to the Free Software 19 * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. 20 * 21 * Copyright (C) IBM Corporation, 2002, 2004 22 * 23 * 2002-Oct Created by Vamsi Krishna S <vamsi_krishna@in.ibm.com> Kernel 24 * Probes initial implementation ( includes suggestions from 25 * Rusty Russell). 26 * 2004-July Suparna Bhattacharya <suparna@in.ibm.com> added jumper probes 27 * interface to access function arguments. 28 * 2005-May Hien Nguyen <hien@us.ibm.com> and Jim Keniston 29 * <jkenisto@us.ibm.com> and Prasanna S Panchamukhi 30 * <prasanna@in.ibm.com> added function-return probes. 31 */ 32#include <linux/compiler.h> 33#include <linux/linkage.h> 34#include <linux/list.h> 35#include <linux/notifier.h> 36#include <linux/smp.h> 37#include <linux/bug.h> 38#include <linux/percpu.h> 39#include <linux/spinlock.h> 40#include <linux/rcupdate.h> 41#include <linux/mutex.h> 42#include <linux/ftrace.h> 43#include <asm/kprobes.h> 44 45#ifdef CONFIG_KPROBES 46 47/* kprobe_status settings */ 48#define KPROBE_HIT_ACTIVE 0x00000001 49#define KPROBE_HIT_SS 0x00000002 50#define KPROBE_REENTER 0x00000004 51#define KPROBE_HIT_SSDONE 0x00000008 52 53#else /* CONFIG_KPROBES */ 54#include <asm-generic/kprobes.h> 55typedef int kprobe_opcode_t; 56struct arch_specific_insn { 57 int dummy; 58}; 59#endif /* CONFIG_KPROBES */ 60 61struct kprobe; 62struct pt_regs; 63struct kretprobe; 64struct kretprobe_instance; 65typedef int (*kprobe_pre_handler_t) (struct kprobe *, struct pt_regs *); 66typedef int (*kprobe_break_handler_t) (struct kprobe *, struct pt_regs *); 67typedef void (*kprobe_post_handler_t) (struct kprobe *, struct pt_regs *, 68 unsigned long flags); 69typedef int (*kprobe_fault_handler_t) (struct kprobe *, struct pt_regs *, 70 int trapnr); 71typedef int (*kretprobe_handler_t) (struct kretprobe_instance *, 72 struct pt_regs *); 73 74struct kprobe { 75 struct hlist_node hlist; 76 77 /* list of kprobes for multi-handler support */ 78 struct list_head list; 79 80 /*count the number of times this probe was temporarily disarmed */ 81 unsigned long nmissed; 82 83 /* location of the probe point */ 84 kprobe_opcode_t *addr; 85 86 /* Allow user to indicate symbol name of the probe point */ 87 const char *symbol_name; 88 89 /* Offset into the symbol */ 90 unsigned int offset; 91 92 /* Called before addr is executed. */ 93 kprobe_pre_handler_t pre_handler; 94 95 /* Called after addr is executed, unless... */ 96 kprobe_post_handler_t post_handler; 97 98 /* 99 * ... called if executing addr causes a fault (eg. page fault). 100 * Return 1 if it handled fault, otherwise kernel will see it. 101 */ 102 kprobe_fault_handler_t fault_handler; 103 104 /* 105 * ... called if breakpoint trap occurs in probe handler. 106 * Return 1 if it handled break, otherwise kernel will see it. 107 */ 108 kprobe_break_handler_t break_handler; 109 110 /* Saved opcode (which has been replaced with breakpoint) */ 111 kprobe_opcode_t opcode; 112 113 /* copy of the original instruction */ 114 struct arch_specific_insn ainsn; 115 116 /* 117 * Indicates various status flags. 118 * Protected by kprobe_mutex after this kprobe is registered. 119 */ 120 u32 flags; 121}; 122 123/* Kprobe status flags */ 124#define KPROBE_FLAG_GONE 1 /* breakpoint has already gone */ 125#define KPROBE_FLAG_DISABLED 2 /* probe is temporarily disabled */ 126#define KPROBE_FLAG_OPTIMIZED 4 /* 127 * probe is really optimized. 128 * NOTE: 129 * this flag is only for optimized_kprobe. 130 */ 131#define KPROBE_FLAG_FTRACE 8 /* probe is using ftrace */ 132 133/* Has this kprobe gone ? */ 134static inline int kprobe_gone(struct kprobe *p) 135{ 136 return p->flags & KPROBE_FLAG_GONE; 137} 138 139/* Is this kprobe disabled ? */ 140static inline int kprobe_disabled(struct kprobe *p) 141{ 142 return p->flags & (KPROBE_FLAG_DISABLED | KPROBE_FLAG_GONE); 143} 144 145/* Is this kprobe really running optimized path ? */ 146static inline int kprobe_optimized(struct kprobe *p) 147{ 148 return p->flags & KPROBE_FLAG_OPTIMIZED; 149} 150 151/* Is this kprobe uses ftrace ? */ 152static inline int kprobe_ftrace(struct kprobe *p) 153{ 154 return p->flags & KPROBE_FLAG_FTRACE; 155} 156 157/* 158 * Special probe type that uses setjmp-longjmp type tricks to resume 159 * execution at a specified entry with a matching prototype corresponding 160 * to the probed function - a trick to enable arguments to become 161 * accessible seamlessly by probe handling logic. 162 * Note: 163 * Because of the way compilers allocate stack space for local variables 164 * etc upfront, regardless of sub-scopes within a function, this mirroring 165 * principle currently works only for probes placed on function entry points. 166 */ 167struct jprobe { 168 struct kprobe kp; 169 void *entry; /* probe handling code to jump to */ 170}; 171 172/* For backward compatibility with old code using JPROBE_ENTRY() */ 173#define JPROBE_ENTRY(handler) (handler) 174 175/* 176 * Function-return probe - 177 * Note: 178 * User needs to provide a handler function, and initialize maxactive. 179 * maxactive - The maximum number of instances of the probed function that 180 * can be active concurrently. 181 * nmissed - tracks the number of times the probed function's return was 182 * ignored, due to maxactive being too low. 183 * 184 */ 185struct kretprobe { 186 struct kprobe kp; 187 kretprobe_handler_t handler; 188 kretprobe_handler_t entry_handler; 189 int maxactive; 190 int nmissed; 191 size_t data_size; 192 struct hlist_head free_instances; 193 raw_spinlock_t lock; 194}; 195 196struct kretprobe_instance { 197 struct hlist_node hlist; 198 struct kretprobe *rp; 199 kprobe_opcode_t *ret_addr; 200 struct task_struct *task; 201 char data[0]; 202}; 203 204struct kretprobe_blackpoint { 205 const char *name; 206 void *addr; 207}; 208 209struct kprobe_blacklist_entry { 210 struct list_head list; 211 unsigned long start_addr; 212 unsigned long end_addr; 213}; 214 215#ifdef CONFIG_KPROBES 216DECLARE_PER_CPU(struct kprobe *, current_kprobe); 217DECLARE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk); 218 219/* 220 * For #ifdef avoidance: 221 */ 222static inline int kprobes_built_in(void) 223{ 224 return 1; 225} 226 227#ifdef CONFIG_KRETPROBES 228extern void arch_prepare_kretprobe(struct kretprobe_instance *ri, 229 struct pt_regs *regs); 230extern int arch_trampoline_kprobe(struct kprobe *p); 231#else /* CONFIG_KRETPROBES */ 232static inline void arch_prepare_kretprobe(struct kretprobe *rp, 233 struct pt_regs *regs) 234{ 235} 236static inline int arch_trampoline_kprobe(struct kprobe *p) 237{ 238 return 0; 239} 240#endif /* CONFIG_KRETPROBES */ 241 242extern struct kretprobe_blackpoint kretprobe_blacklist[]; 243 244static inline void kretprobe_assert(struct kretprobe_instance *ri, 245 unsigned long orig_ret_address, unsigned long trampoline_address) 246{ 247 if (!orig_ret_address || (orig_ret_address == trampoline_address)) { 248 printk("kretprobe BUG!: Processing kretprobe %p @ %p\n", 249 ri->rp, ri->rp->kp.addr); 250 BUG(); 251 } 252} 253 254#ifdef CONFIG_KPROBES_SANITY_TEST 255extern int init_test_probes(void); 256#else 257static inline int init_test_probes(void) 258{ 259 return 0; 260} 261#endif /* CONFIG_KPROBES_SANITY_TEST */ 262 263extern int arch_prepare_kprobe(struct kprobe *p); 264extern void arch_arm_kprobe(struct kprobe *p); 265extern void arch_disarm_kprobe(struct kprobe *p); 266extern int arch_init_kprobes(void); 267extern void show_registers(struct pt_regs *regs); 268extern void kprobes_inc_nmissed_count(struct kprobe *p); 269extern bool arch_within_kprobe_blacklist(unsigned long addr); 270extern bool arch_kprobe_on_func_entry(unsigned long offset); 271extern bool kprobe_on_func_entry(kprobe_opcode_t *addr, const char *sym, unsigned long offset); 272 273extern bool within_kprobe_blacklist(unsigned long addr); 274 275struct kprobe_insn_cache { 276 struct mutex mutex; 277 void *(*alloc)(void); /* allocate insn page */ 278 void (*free)(void *); /* free insn page */ 279 struct list_head pages; /* list of kprobe_insn_page */ 280 size_t insn_size; /* size of instruction slot */ 281 int nr_garbage; 282}; 283 284#ifdef __ARCH_WANT_KPROBES_INSN_SLOT 285extern kprobe_opcode_t *__get_insn_slot(struct kprobe_insn_cache *c); 286extern void __free_insn_slot(struct kprobe_insn_cache *c, 287 kprobe_opcode_t *slot, int dirty); 288/* sleep-less address checking routine */ 289extern bool __is_insn_slot_addr(struct kprobe_insn_cache *c, 290 unsigned long addr); 291 292#define DEFINE_INSN_CACHE_OPS(__name) \ 293extern struct kprobe_insn_cache kprobe_##__name##_slots; \ 294 \ 295static inline kprobe_opcode_t *get_##__name##_slot(void) \ 296{ \ 297 return __get_insn_slot(&kprobe_##__name##_slots); \ 298} \ 299 \ 300static inline void free_##__name##_slot(kprobe_opcode_t *slot, int dirty)\ 301{ \ 302 __free_insn_slot(&kprobe_##__name##_slots, slot, dirty); \ 303} \ 304 \ 305static inline bool is_kprobe_##__name##_slot(unsigned long addr) \ 306{ \ 307 return __is_insn_slot_addr(&kprobe_##__name##_slots, addr); \ 308} 309#else /* __ARCH_WANT_KPROBES_INSN_SLOT */ 310#define DEFINE_INSN_CACHE_OPS(__name) \ 311static inline bool is_kprobe_##__name##_slot(unsigned long addr) \ 312{ \ 313 return 0; \ 314} 315#endif 316 317DEFINE_INSN_CACHE_OPS(insn); 318 319#ifdef CONFIG_OPTPROBES 320/* 321 * Internal structure for direct jump optimized probe 322 */ 323struct optimized_kprobe { 324 struct kprobe kp; 325 struct list_head list; /* list for optimizing queue */ 326 struct arch_optimized_insn optinsn; 327}; 328 329/* Architecture dependent functions for direct jump optimization */ 330extern int arch_prepared_optinsn(struct arch_optimized_insn *optinsn); 331extern int arch_check_optimized_kprobe(struct optimized_kprobe *op); 332extern int arch_prepare_optimized_kprobe(struct optimized_kprobe *op, 333 struct kprobe *orig); 334extern void arch_remove_optimized_kprobe(struct optimized_kprobe *op); 335extern void arch_optimize_kprobes(struct list_head *oplist); 336extern void arch_unoptimize_kprobes(struct list_head *oplist, 337 struct list_head *done_list); 338extern void arch_unoptimize_kprobe(struct optimized_kprobe *op); 339extern int arch_within_optimized_kprobe(struct optimized_kprobe *op, 340 unsigned long addr); 341 342extern void opt_pre_handler(struct kprobe *p, struct pt_regs *regs); 343 344DEFINE_INSN_CACHE_OPS(optinsn); 345 346#ifdef CONFIG_SYSCTL 347extern int sysctl_kprobes_optimization; 348extern int proc_kprobes_optimization_handler(struct ctl_table *table, 349 int write, void __user *buffer, 350 size_t *length, loff_t *ppos); 351#endif 352extern void wait_for_kprobe_optimizer(void); 353#else 354static inline void wait_for_kprobe_optimizer(void) { } 355#endif /* CONFIG_OPTPROBES */ 356#ifdef CONFIG_KPROBES_ON_FTRACE 357extern void kprobe_ftrace_handler(unsigned long ip, unsigned long parent_ip, 358 struct ftrace_ops *ops, struct pt_regs *regs); 359extern int arch_prepare_kprobe_ftrace(struct kprobe *p); 360#endif 361 362int arch_check_ftrace_location(struct kprobe *p); 363 364/* Get the kprobe at this addr (if any) - called with preemption disabled */ 365struct kprobe *get_kprobe(void *addr); 366void kretprobe_hash_lock(struct task_struct *tsk, 367 struct hlist_head **head, unsigned long *flags); 368void kretprobe_hash_unlock(struct task_struct *tsk, unsigned long *flags); 369struct hlist_head * kretprobe_inst_table_head(struct task_struct *tsk); 370 371/* kprobe_running() will just return the current_kprobe on this CPU */ 372static inline struct kprobe *kprobe_running(void) 373{ 374 return (__this_cpu_read(current_kprobe)); 375} 376 377static inline void reset_current_kprobe(void) 378{ 379 __this_cpu_write(current_kprobe, NULL); 380} 381 382static inline struct kprobe_ctlblk *get_kprobe_ctlblk(void) 383{ 384 return this_cpu_ptr(&kprobe_ctlblk); 385} 386 387kprobe_opcode_t *kprobe_lookup_name(const char *name, unsigned int offset); 388int register_kprobe(struct kprobe *p); 389void unregister_kprobe(struct kprobe *p); 390int register_kprobes(struct kprobe **kps, int num); 391void unregister_kprobes(struct kprobe **kps, int num); 392int setjmp_pre_handler(struct kprobe *, struct pt_regs *); 393int longjmp_break_handler(struct kprobe *, struct pt_regs *); 394int register_jprobe(struct jprobe *p); 395void unregister_jprobe(struct jprobe *p); 396int register_jprobes(struct jprobe **jps, int num); 397void unregister_jprobes(struct jprobe **jps, int num); 398void jprobe_return(void); 399unsigned long arch_deref_entry_point(void *); 400 401int register_kretprobe(struct kretprobe *rp); 402void unregister_kretprobe(struct kretprobe *rp); 403int register_kretprobes(struct kretprobe **rps, int num); 404void unregister_kretprobes(struct kretprobe **rps, int num); 405 406void kprobe_flush_task(struct task_struct *tk); 407void recycle_rp_inst(struct kretprobe_instance *ri, struct hlist_head *head); 408 409int disable_kprobe(struct kprobe *kp); 410int enable_kprobe(struct kprobe *kp); 411 412void dump_kprobe(struct kprobe *kp); 413 414#else /* !CONFIG_KPROBES: */ 415 416static inline int kprobes_built_in(void) 417{ 418 return 0; 419} 420static inline int kprobe_fault_handler(struct pt_regs *regs, int trapnr) 421{ 422 return 0; 423} 424static inline struct kprobe *get_kprobe(void *addr) 425{ 426 return NULL; 427} 428static inline struct kprobe *kprobe_running(void) 429{ 430 return NULL; 431} 432static inline int register_kprobe(struct kprobe *p) 433{ 434 return -ENOSYS; 435} 436static inline int register_kprobes(struct kprobe **kps, int num) 437{ 438 return -ENOSYS; 439} 440static inline void unregister_kprobe(struct kprobe *p) 441{ 442} 443static inline void unregister_kprobes(struct kprobe **kps, int num) 444{ 445} 446static inline int register_jprobe(struct jprobe *p) 447{ 448 return -ENOSYS; 449} 450static inline int register_jprobes(struct jprobe **jps, int num) 451{ 452 return -ENOSYS; 453} 454static inline void unregister_jprobe(struct jprobe *p) 455{ 456} 457static inline void unregister_jprobes(struct jprobe **jps, int num) 458{ 459} 460static inline void jprobe_return(void) 461{ 462} 463static inline int register_kretprobe(struct kretprobe *rp) 464{ 465 return -ENOSYS; 466} 467static inline int register_kretprobes(struct kretprobe **rps, int num) 468{ 469 return -ENOSYS; 470} 471static inline void unregister_kretprobe(struct kretprobe *rp) 472{ 473} 474static inline void unregister_kretprobes(struct kretprobe **rps, int num) 475{ 476} 477static inline void kprobe_flush_task(struct task_struct *tk) 478{ 479} 480static inline int disable_kprobe(struct kprobe *kp) 481{ 482 return -ENOSYS; 483} 484static inline int enable_kprobe(struct kprobe *kp) 485{ 486 return -ENOSYS; 487} 488#endif /* CONFIG_KPROBES */ 489static inline int disable_kretprobe(struct kretprobe *rp) 490{ 491 return disable_kprobe(&rp->kp); 492} 493static inline int enable_kretprobe(struct kretprobe *rp) 494{ 495 return enable_kprobe(&rp->kp); 496} 497static inline int disable_jprobe(struct jprobe *jp) 498{ 499 return disable_kprobe(&jp->kp); 500} 501static inline int enable_jprobe(struct jprobe *jp) 502{ 503 return enable_kprobe(&jp->kp); 504} 505 506#ifndef CONFIG_KPROBES 507static inline bool is_kprobe_insn_slot(unsigned long addr) 508{ 509 return false; 510} 511#endif 512#ifndef CONFIG_OPTPROBES 513static inline bool is_kprobe_optinsn_slot(unsigned long addr) 514{ 515 return false; 516} 517#endif 518 519#endif /* _LINUX_KPROBES_H */