kprobes.c 28.5 KB
Newer Older
L
Linus Torvalds 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29
/*
 *  Kernel Probes (KProbes)
 *  kernel/kprobes.c
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License as published by
 * the Free Software Foundation; either version 2 of the License, or
 * (at your option) any later version.
 *
 * This program is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program; if not, write to the Free Software
 * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
 *
 * Copyright (C) IBM Corporation, 2002, 2004
 *
 * 2002-Oct	Created by Vamsi Krishna S <vamsi_krishna@in.ibm.com> Kernel
 *		Probes initial implementation (includes suggestions from
 *		Rusty Russell).
 * 2004-Aug	Updated by Prasanna S Panchamukhi <prasanna@in.ibm.com> with
 *		hlists and exceptions notifier as suggested by Andi Kleen.
 * 2004-July	Suparna Bhattacharya <suparna@in.ibm.com> added jumper probes
 *		interface to access function arguments.
 * 2004-Sep	Prasanna S Panchamukhi <prasanna@in.ibm.com> Changed Kprobes
 *		exceptions notifier to be first on the priority list.
30 31 32
 * 2005-May	Hien Nguyen <hien@us.ibm.com>, Jim Keniston
 *		<jkenisto@us.ibm.com> and Prasanna S Panchamukhi
 *		<prasanna@in.ibm.com> added function-return probes.
L
Linus Torvalds 已提交
33 34 35 36
 */
#include <linux/kprobes.h>
#include <linux/hash.h>
#include <linux/init.h>
T
Tim Schmielau 已提交
37
#include <linux/slab.h>
R
Randy Dunlap 已提交
38
#include <linux/stddef.h>
L
Linus Torvalds 已提交
39
#include <linux/module.h>
40
#include <linux/moduleloader.h>
41
#include <linux/kallsyms.h>
42
#include <linux/freezer.h>
43 44
#include <linux/seq_file.h>
#include <linux/debugfs.h>
45
#include <linux/kdebug.h>
46

47
#include <asm-generic/sections.h>
L
Linus Torvalds 已提交
48 49
#include <asm/cacheflush.h>
#include <asm/errno.h>
50
#include <asm/uaccess.h>
L
Linus Torvalds 已提交
51 52 53 54

#define KPROBE_HASH_BITS 6
#define KPROBE_TABLE_SIZE (1 << KPROBE_HASH_BITS)

55 56 57 58 59 60 61 62 63 64

/*
 * Some oddball architectures like 64bit powerpc have function descriptors
 * so this must be overridable.
 */
#ifndef kprobe_lookup_name
#define kprobe_lookup_name(name, addr) \
	addr = ((kprobe_opcode_t *)(kallsyms_lookup_name(name)))
#endif

L
Linus Torvalds 已提交
65
static struct hlist_head kprobe_table[KPROBE_TABLE_SIZE];
66
static struct hlist_head kretprobe_inst_table[KPROBE_TABLE_SIZE];
L
Linus Torvalds 已提交
67

68 69 70
/* NOTE: change this value only with kprobe_mutex held */
static bool kprobe_enabled;

I
Ingo Molnar 已提交
71
DEFINE_MUTEX(kprobe_mutex);		/* Protects kprobe_table */
72
DEFINE_SPINLOCK(kretprobe_lock);	/* Protects kretprobe_inst_table */
73
static DEFINE_PER_CPU(struct kprobe *, kprobe_instance) = NULL;
L
Linus Torvalds 已提交
74

75 76 77 78 79 80 81 82 83 84 85 86
/*
 * Normally, functions that we'd want to prohibit kprobes in, are marked
 * __kprobes. But, there are cases where such functions already belong to
 * a different section (__sched for preempt_schedule)
 *
 * For such cases, we now have a blacklist
 */
struct kprobe_blackpoint kprobe_blacklist[] = {
	{"preempt_schedule",},
	{NULL}    /* Terminator */
};

87
#ifdef __ARCH_WANT_KPROBES_INSN_SLOT
88 89 90 91 92 93 94 95 96 97 98 99 100
/*
 * kprobe->ainsn.insn points to the copy of the instruction to be
 * single-stepped. x86_64, POWER4 and above have no-exec support and
 * stepping on the instruction on a vmalloced/kmalloced/data page
 * is a recipe for disaster
 */
#define INSNS_PER_PAGE	(PAGE_SIZE/(MAX_INSN_SIZE * sizeof(kprobe_opcode_t)))

struct kprobe_insn_page {
	struct hlist_node hlist;
	kprobe_opcode_t *insns;		/* Page of instruction slots */
	char slot_used[INSNS_PER_PAGE];
	int nused;
101
	int ngarbage;
102 103
};

104 105 106 107 108 109
enum kprobe_slot_state {
	SLOT_CLEAN = 0,
	SLOT_DIRTY = 1,
	SLOT_USED = 2,
};

110
static struct hlist_head kprobe_insn_pages;
111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136
static int kprobe_garbage_slots;
static int collect_garbage_slots(void);

static int __kprobes check_safety(void)
{
	int ret = 0;
#if defined(CONFIG_PREEMPT) && defined(CONFIG_PM)
	ret = freeze_processes();
	if (ret == 0) {
		struct task_struct *p, *q;
		do_each_thread(p, q) {
			if (p != current && p->state == TASK_RUNNING &&
			    p->pid != 0) {
				printk("Check failed: %s is running\n",p->comm);
				ret = -1;
				goto loop_end;
			}
		} while_each_thread(p, q);
	}
loop_end:
	thaw_processes();
#else
	synchronize_sched();
#endif
	return ret;
}
137 138 139 140 141

/**
 * get_insn_slot() - Find a slot on an executable page for an instruction.
 * We allocate an executable page if there's no room on existing ones.
 */
142
kprobe_opcode_t __kprobes *get_insn_slot(void)
143 144 145 146
{
	struct kprobe_insn_page *kip;
	struct hlist_node *pos;

147
 retry:
148
	hlist_for_each_entry(kip, pos, &kprobe_insn_pages, hlist) {
149 150 151
		if (kip->nused < INSNS_PER_PAGE) {
			int i;
			for (i = 0; i < INSNS_PER_PAGE; i++) {
152 153
				if (kip->slot_used[i] == SLOT_CLEAN) {
					kip->slot_used[i] = SLOT_USED;
154 155 156 157 158 159 160 161 162
					kip->nused++;
					return kip->insns + (i * MAX_INSN_SIZE);
				}
			}
			/* Surprise!  No unused slots.  Fix kip->nused. */
			kip->nused = INSNS_PER_PAGE;
		}
	}

163 164 165 166 167
	/* If there are any garbage slots, collect it and try again. */
	if (kprobe_garbage_slots && collect_garbage_slots() == 0) {
		goto retry;
	}
	/* All out of space.  Need to allocate a new page. Use slot 0. */
168
	kip = kmalloc(sizeof(struct kprobe_insn_page), GFP_KERNEL);
169
	if (!kip)
170 171 172 173 174 175 176 177 178 179 180 181 182 183
		return NULL;

	/*
	 * Use module_alloc so this page is within +/- 2GB of where the
	 * kernel image and loaded module images reside. This is required
	 * so x86_64 can correctly handle the %rip-relative fixups.
	 */
	kip->insns = module_alloc(PAGE_SIZE);
	if (!kip->insns) {
		kfree(kip);
		return NULL;
	}
	INIT_HLIST_NODE(&kip->hlist);
	hlist_add_head(&kip->hlist, &kprobe_insn_pages);
184 185
	memset(kip->slot_used, SLOT_CLEAN, INSNS_PER_PAGE);
	kip->slot_used[0] = SLOT_USED;
186
	kip->nused = 1;
187
	kip->ngarbage = 0;
188 189 190
	return kip->insns;
}

191 192 193
/* Return 1 if all garbages are collected, otherwise 0. */
static int __kprobes collect_one_slot(struct kprobe_insn_page *kip, int idx)
{
194
	kip->slot_used[idx] = SLOT_CLEAN;
195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225
	kip->nused--;
	if (kip->nused == 0) {
		/*
		 * Page is no longer in use.  Free it unless
		 * it's the last one.  We keep the last one
		 * so as not to have to set it up again the
		 * next time somebody inserts a probe.
		 */
		hlist_del(&kip->hlist);
		if (hlist_empty(&kprobe_insn_pages)) {
			INIT_HLIST_NODE(&kip->hlist);
			hlist_add_head(&kip->hlist,
				       &kprobe_insn_pages);
		} else {
			module_free(NULL, kip->insns);
			kfree(kip);
		}
		return 1;
	}
	return 0;
}

static int __kprobes collect_garbage_slots(void)
{
	struct kprobe_insn_page *kip;
	struct hlist_node *pos, *next;

	/* Ensure no-one is preepmted on the garbages */
	if (check_safety() != 0)
		return -EAGAIN;

226
	hlist_for_each_entry_safe(kip, pos, next, &kprobe_insn_pages, hlist) {
227 228 229 230 231
		int i;
		if (kip->ngarbage == 0)
			continue;
		kip->ngarbage = 0;	/* we will collect all garbages */
		for (i = 0; i < INSNS_PER_PAGE; i++) {
232
			if (kip->slot_used[i] == SLOT_DIRTY &&
233 234 235 236 237 238 239 240 241
			    collect_one_slot(kip, i))
				break;
		}
	}
	kprobe_garbage_slots = 0;
	return 0;
}

void __kprobes free_insn_slot(kprobe_opcode_t * slot, int dirty)
242 243 244 245
{
	struct kprobe_insn_page *kip;
	struct hlist_node *pos;

246
	hlist_for_each_entry(kip, pos, &kprobe_insn_pages, hlist) {
247 248 249
		if (kip->insns <= slot &&
		    slot < kip->insns + (INSNS_PER_PAGE * MAX_INSN_SIZE)) {
			int i = (slot - kip->insns) / MAX_INSN_SIZE;
250
			if (dirty) {
251
				kip->slot_used[i] = SLOT_DIRTY;
252 253 254
				kip->ngarbage++;
			} else {
				collect_one_slot(kip, i);
255
			}
256
			break;
257 258
		}
	}
259 260

	if (dirty && ++kprobe_garbage_slots > INSNS_PER_PAGE)
261
		collect_garbage_slots();
262
}
263
#endif
264

265 266 267 268 269 270 271 272 273 274 275
/* We have preemption disabled.. so it is safe to use __ versions */
static inline void set_kprobe_instance(struct kprobe *kp)
{
	__get_cpu_var(kprobe_instance) = kp;
}

static inline void reset_kprobe_instance(void)
{
	__get_cpu_var(kprobe_instance) = NULL;
}

276 277
/*
 * This routine is called either:
278
 * 	- under the kprobe_mutex - during kprobe_[un]register()
279
 * 				OR
280
 * 	- with preemption disabled - from arch/xxx/kernel/kprobes.c
281
 */
282
struct kprobe __kprobes *get_kprobe(void *addr)
L
Linus Torvalds 已提交
283 284 285
{
	struct hlist_head *head;
	struct hlist_node *node;
286
	struct kprobe *p;
L
Linus Torvalds 已提交
287 288

	head = &kprobe_table[hash_ptr(addr, KPROBE_HASH_BITS)];
289
	hlist_for_each_entry_rcu(p, node, head, hlist) {
L
Linus Torvalds 已提交
290 291 292 293 294 295
		if (p->addr == addr)
			return p;
	}
	return NULL;
}

296 297 298 299
/*
 * Aggregate handlers for multiple kprobes support - these handlers
 * take care of invoking the individual kprobe handlers on p->list
 */
300
static int __kprobes aggr_pre_handler(struct kprobe *p, struct pt_regs *regs)
301 302 303
{
	struct kprobe *kp;

304
	list_for_each_entry_rcu(kp, &p->list, list) {
305
		if (kp->pre_handler) {
306
			set_kprobe_instance(kp);
307 308
			if (kp->pre_handler(kp, regs))
				return 1;
309
		}
310
		reset_kprobe_instance();
311 312 313 314
	}
	return 0;
}

315 316
static void __kprobes aggr_post_handler(struct kprobe *p, struct pt_regs *regs,
					unsigned long flags)
317 318 319
{
	struct kprobe *kp;

320
	list_for_each_entry_rcu(kp, &p->list, list) {
321
		if (kp->post_handler) {
322
			set_kprobe_instance(kp);
323
			kp->post_handler(kp, regs, flags);
324
			reset_kprobe_instance();
325 326 327 328
		}
	}
}

329 330
static int __kprobes aggr_fault_handler(struct kprobe *p, struct pt_regs *regs,
					int trapnr)
331
{
332 333
	struct kprobe *cur = __get_cpu_var(kprobe_instance);

334 335 336 337
	/*
	 * if we faulted "during" the execution of a user specified
	 * probe handler, invoke just that probe's fault handler
	 */
338 339
	if (cur && cur->fault_handler) {
		if (cur->fault_handler(cur, regs, trapnr))
340 341 342 343 344
			return 1;
	}
	return 0;
}

345
static int __kprobes aggr_break_handler(struct kprobe *p, struct pt_regs *regs)
346
{
347 348 349 350 351 352
	struct kprobe *cur = __get_cpu_var(kprobe_instance);
	int ret = 0;

	if (cur && cur->break_handler) {
		if (cur->break_handler(cur, regs))
			ret = 1;
353
	}
354 355
	reset_kprobe_instance();
	return ret;
356 357
}

358 359 360 361 362 363 364 365 366 367 368 369 370
/* Walks the list and increments nmissed count for multiprobe case */
void __kprobes kprobes_inc_nmissed_count(struct kprobe *p)
{
	struct kprobe *kp;
	if (p->pre_handler != aggr_pre_handler) {
		p->nmissed++;
	} else {
		list_for_each_entry_rcu(kp, &p->list, list)
			kp->nmissed++;
	}
	return;
}

371
/* Called with kretprobe_lock held */
372 373
void __kprobes recycle_rp_inst(struct kretprobe_instance *ri,
				struct hlist_head *head)
374 375 376 377 378 379 380 381 382 383 384
{
	/* remove rp inst off the rprobe_inst_table */
	hlist_del(&ri->hlist);
	if (ri->rp) {
		/* remove rp inst off the used list */
		hlist_del(&ri->uflist);
		/* put rp inst back onto the free list */
		INIT_HLIST_NODE(&ri->uflist);
		hlist_add_head(&ri->uflist, &ri->rp->free_instances);
	} else
		/* Unregistering */
385
		hlist_add_head(&ri->hlist, head);
386 387
}

388
struct hlist_head __kprobes *kretprobe_inst_table_head(struct task_struct *tsk)
389 390 391 392 393
{
	return &kretprobe_inst_table[hash_ptr(tsk, KPROBE_HASH_BITS)];
}

/*
394 395 396 397
 * This function is called from finish_task_switch when task tk becomes dead,
 * so that we can recycle any function-return probe instances associated
 * with this task. These left over instances represent probed functions
 * that have been called but will never return.
398
 */
399
void __kprobes kprobe_flush_task(struct task_struct *tk)
400
{
B
bibo,mao 已提交
401
	struct kretprobe_instance *ri;
402
	struct hlist_head *head, empty_rp;
403
	struct hlist_node *node, *tmp;
404
	unsigned long flags = 0;
405

406
	INIT_HLIST_HEAD(&empty_rp);
407
	spin_lock_irqsave(&kretprobe_lock, flags);
B
bibo,mao 已提交
408 409 410
	head = kretprobe_inst_table_head(tk);
	hlist_for_each_entry_safe(ri, node, tmp, head, hlist) {
		if (ri->task == tk)
411
			recycle_rp_inst(ri, &empty_rp);
B
bibo,mao 已提交
412
	}
413
	spin_unlock_irqrestore(&kretprobe_lock, flags);
414 415 416 417 418

	hlist_for_each_entry_safe(ri, node, tmp, &empty_rp, hlist) {
		hlist_del(&ri->hlist);
		kfree(ri);
	}
419 420 421 422 423
}

static inline void free_rp_inst(struct kretprobe *rp)
{
	struct kretprobe_instance *ri;
424 425 426
	struct hlist_node *pos, *next;

	hlist_for_each_entry_safe(ri, pos, next, &rp->free_instances, uflist) {
427 428 429 430 431
		hlist_del(&ri->uflist);
		kfree(ri);
	}
}

432 433 434 435 436 437 438 439 440 441 442 443 444
/*
 * Keep all fields in the kprobe consistent
 */
static inline void copy_kprobe(struct kprobe *old_p, struct kprobe *p)
{
	memcpy(&p->opcode, &old_p->opcode, sizeof(kprobe_opcode_t));
	memcpy(&p->ainsn, &old_p->ainsn, sizeof(struct arch_specific_insn));
}

/*
* Add the new probe to old_p->list. Fail if this is the
* second jprobe at the address - two jprobes can't coexist
*/
445
static int __kprobes add_new_kprobe(struct kprobe *old_p, struct kprobe *p)
446 447
{
	if (p->break_handler) {
448 449
		if (old_p->break_handler)
			return -EEXIST;
450
		list_add_tail_rcu(&p->list, &old_p->list);
451
		old_p->break_handler = aggr_break_handler;
452
	} else
453
		list_add_rcu(&p->list, &old_p->list);
454 455
	if (p->post_handler && !old_p->post_handler)
		old_p->post_handler = aggr_post_handler;
456 457 458
	return 0;
}

459 460 461 462 463 464
/*
 * Fill in the required fields of the "manager kprobe". Replace the
 * earlier kprobe in the hlist with the manager kprobe
 */
static inline void add_aggr_kprobe(struct kprobe *ap, struct kprobe *p)
{
465
	copy_kprobe(p, ap);
466
	flush_insn_slot(ap);
467 468 469
	ap->addr = p->addr;
	ap->pre_handler = aggr_pre_handler;
	ap->fault_handler = aggr_fault_handler;
470 471 472 473
	if (p->post_handler)
		ap->post_handler = aggr_post_handler;
	if (p->break_handler)
		ap->break_handler = aggr_break_handler;
474 475

	INIT_LIST_HEAD(&ap->list);
476
	list_add_rcu(&p->list, &ap->list);
477

478
	hlist_replace_rcu(&p->hlist, &ap->hlist);
479 480 481 482 483 484
}

/*
 * This is the second or subsequent kprobe at the address - handle
 * the intricacies
 */
485 486
static int __kprobes register_aggr_kprobe(struct kprobe *old_p,
					  struct kprobe *p)
487 488 489 490
{
	int ret = 0;
	struct kprobe *ap;

491 492 493
	if (old_p->pre_handler == aggr_pre_handler) {
		copy_kprobe(old_p, p);
		ret = add_new_kprobe(old_p, p);
494
	} else {
495
		ap = kzalloc(sizeof(struct kprobe), GFP_KERNEL);
496 497 498
		if (!ap)
			return -ENOMEM;
		add_aggr_kprobe(ap, old_p);
499 500
		copy_kprobe(ap, p);
		ret = add_new_kprobe(ap, p);
501 502 503 504
	}
	return ret;
}

505 506
static int __kprobes in_kprobes_functions(unsigned long addr)
{
507 508
	struct kprobe_blackpoint *kb;

509 510
	if (addr >= (unsigned long)__kprobes_text_start &&
	    addr < (unsigned long)__kprobes_text_end)
511
		return -EINVAL;
512 513 514 515 516 517 518 519 520 521 522
	/*
	 * If there exists a kprobe_blacklist, verify and
	 * fail any probe registration in the prohibited area
	 */
	for (kb = kprobe_blacklist; kb->name != NULL; kb++) {
		if (kb->start_addr) {
			if (addr >= kb->start_addr &&
			    addr < (kb->start_addr + kb->range))
				return -EINVAL;
		}
	}
523 524 525
	return 0;
}

526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543
/*
 * If we have a symbol_name argument, look it up and add the offset field
 * to it. This way, we can specify a relative address to a symbol.
 */
static kprobe_opcode_t __kprobes *kprobe_addr(struct kprobe *p)
{
	kprobe_opcode_t *addr = p->addr;
	if (p->symbol_name) {
		if (addr)
			return NULL;
		kprobe_lookup_name(p->symbol_name, addr);
	}

	if (!addr)
		return NULL;
	return (kprobe_opcode_t *)(((char *)addr) + p->offset);
}

544 545
static int __kprobes __register_kprobe(struct kprobe *p,
	unsigned long called_from)
L
Linus Torvalds 已提交
546 547
{
	int ret = 0;
548
	struct kprobe *old_p;
549
	struct module *probed_mod;
550
	kprobe_opcode_t *addr;
551

552 553
	addr = kprobe_addr(p);
	if (!addr)
554
		return -EINVAL;
555
	p->addr = addr;
556

557 558
	if (!kernel_text_address((unsigned long) p->addr) ||
	    in_kprobes_functions((unsigned long) p->addr))
559 560
		return -EINVAL;

561
	p->mod_refcounted = 0;
562 563 564 565 566 567

	/*
	 * Check if are we probing a module.
	 */
	probed_mod = module_text_address((unsigned long) p->addr);
	if (probed_mod) {
568
		struct module *calling_mod = module_text_address(called_from);
569 570 571 572
		/*
		 * We must allow modules to probe themself and in this case
		 * avoid incrementing the module refcount, so as to allow
		 * unloading of self probing modules.
573
		 */
574
		if (calling_mod && calling_mod != probed_mod) {
575 576 577 578 579 580
			if (unlikely(!try_module_get(probed_mod)))
				return -EINVAL;
			p->mod_refcounted = 1;
		} else
			probed_mod = NULL;
	}
L
Linus Torvalds 已提交
581

582
	p->nmissed = 0;
583
	INIT_LIST_HEAD(&p->list);
I
Ingo Molnar 已提交
584
	mutex_lock(&kprobe_mutex);
585 586 587
	old_p = get_kprobe(p->addr);
	if (old_p) {
		ret = register_aggr_kprobe(old_p, p);
L
Linus Torvalds 已提交
588 589 590
		goto out;
	}

591 592
	ret = arch_prepare_kprobe(p);
	if (ret)
593 594
		goto out;

595
	INIT_HLIST_NODE(&p->hlist);
596
	hlist_add_head_rcu(&p->hlist,
L
Linus Torvalds 已提交
597 598
		       &kprobe_table[hash_ptr(p->addr, KPROBE_HASH_BITS)]);

599
	if (kprobe_enabled)
600
		arch_arm_kprobe(p);
601

L
Linus Torvalds 已提交
602
out:
I
Ingo Molnar 已提交
603
	mutex_unlock(&kprobe_mutex);
604

605 606
	if (ret && probed_mod)
		module_put(probed_mod);
L
Linus Torvalds 已提交
607 608 609
	return ret;
}

610 611 612 613
/*
 * Unregister a kprobe without a scheduler synchronization.
 */
static int __kprobes __unregister_kprobe_top(struct kprobe *p)
L
Linus Torvalds 已提交
614
{
615
	struct kprobe *old_p, *list_p;
616 617

	old_p = get_kprobe(p->addr);
618 619 620
	if (unlikely(!old_p))
		return -EINVAL;

621 622 623 624 625
	if (p != old_p) {
		list_for_each_entry_rcu(list_p, &old_p->list, list)
			if (list_p == p)
			/* kprobe p is a valid probe */
				goto valid_p;
626
		return -EINVAL;
627 628
	}
valid_p:
629 630
	if (old_p == p ||
	    (old_p->pre_handler == aggr_pre_handler &&
631
	     list_is_singular(&old_p->list))) {
632 633 634 635 636 637 638
		/*
		 * Only probe on the hash list. Disarm only if kprobes are
		 * enabled - otherwise, the breakpoint would already have
		 * been removed. We save on flushing icache.
		 */
		if (kprobe_enabled)
			arch_disarm_kprobe(p);
639 640
		hlist_del_rcu(&old_p->hlist);
	} else {
641 642 643 644 645 646 647 648 649 650
		if (p->break_handler)
			old_p->break_handler = NULL;
		if (p->post_handler) {
			list_for_each_entry_rcu(list_p, &old_p->list, list) {
				if ((list_p != p) && (list_p->post_handler))
					goto noclean;
			}
			old_p->post_handler = NULL;
		}
noclean:
651 652
		list_del_rcu(&p->list);
	}
653 654
	return 0;
}
655

656 657 658 659
static void __kprobes __unregister_kprobe_bottom(struct kprobe *p)
{
	struct module *mod;
	struct kprobe *old_p;
660

661 662 663 664 665
	if (p->mod_refcounted) {
		mod = module_text_address((unsigned long)p->addr);
		if (mod)
			module_put(mod);
	}
666

667 668 669 670 671
	if (list_empty(&p->list) || list_is_singular(&p->list)) {
		if (!list_empty(&p->list)) {
			/* "p" is the last child of an aggr_kprobe */
			old_p = list_entry(p->list.next, struct kprobe, list);
			list_del(&p->list);
672
			kfree(old_p);
673
		}
674
		arch_remove_kprobe(p);
675 676 677 678 679 680 681 682 683 684 685 686 687 688 689
	}
}

static int __register_kprobes(struct kprobe **kps, int num,
	unsigned long called_from)
{
	int i, ret = 0;

	if (num <= 0)
		return -EINVAL;
	for (i = 0; i < num; i++) {
		ret = __register_kprobe(kps[i], called_from);
		if (ret < 0 && i > 0) {
			unregister_kprobes(kps, i);
			break;
690
		}
691
	}
692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730
	return ret;
}

/*
 * Registration and unregistration functions for kprobe.
 */
int __kprobes register_kprobe(struct kprobe *p)
{
	return __register_kprobes(&p, 1,
				  (unsigned long)__builtin_return_address(0));
}

void __kprobes unregister_kprobe(struct kprobe *p)
{
	unregister_kprobes(&p, 1);
}

int __kprobes register_kprobes(struct kprobe **kps, int num)
{
	return __register_kprobes(kps, num,
				  (unsigned long)__builtin_return_address(0));
}

void __kprobes unregister_kprobes(struct kprobe **kps, int num)
{
	int i;

	if (num <= 0)
		return;
	mutex_lock(&kprobe_mutex);
	for (i = 0; i < num; i++)
		if (__unregister_kprobe_top(kps[i]) < 0)
			kps[i]->addr = NULL;
	mutex_unlock(&kprobe_mutex);

	synchronize_sched();
	for (i = 0; i < num; i++)
		if (kps[i]->addr)
			__unregister_kprobe_bottom(kps[i]);
L
Linus Torvalds 已提交
731 732 733
}

static struct notifier_block kprobe_exceptions_nb = {
734 735 736 737
	.notifier_call = kprobe_exceptions_notify,
	.priority = 0x7fffffff /* we need to be notified first */
};

738 739 740 741
unsigned long __weak arch_deref_entry_point(void *entry)
{
	return (unsigned long)entry;
}
L
Linus Torvalds 已提交
742

743
int __kprobes register_jprobe(struct jprobe *jp)
L
Linus Torvalds 已提交
744
{
745 746 747 748 749
	unsigned long addr = arch_deref_entry_point(jp->entry);

	if (!kernel_text_address(addr))
		return -EINVAL;

L
Linus Torvalds 已提交
750 751 752 753
	/* Todo: Verify probepoint is a function entry point */
	jp->kp.pre_handler = setjmp_pre_handler;
	jp->kp.break_handler = longjmp_break_handler;

754 755
	return __register_kprobe(&jp->kp,
		(unsigned long)__builtin_return_address(0));
L
Linus Torvalds 已提交
756 757
}

758
void __kprobes unregister_jprobe(struct jprobe *jp)
L
Linus Torvalds 已提交
759 760 761 762
{
	unregister_kprobe(&jp->kp);
}

763
#ifdef CONFIG_KRETPROBES
764 765 766 767 768 769 770 771 772 773 774 775
/*
 * This kprobe pre_handler is registered with every kretprobe. When probe
 * hits it will set up the return probe.
 */
static int __kprobes pre_handler_kretprobe(struct kprobe *p,
					   struct pt_regs *regs)
{
	struct kretprobe *rp = container_of(p, struct kretprobe, kp);
	unsigned long flags = 0;

	/*TODO: consider to only swap the RA after the last pre_handler fired */
	spin_lock_irqsave(&kretprobe_lock, flags);
776 777 778 779 780 781 782
	if (!hlist_empty(&rp->free_instances)) {
		struct kretprobe_instance *ri;

		ri = hlist_entry(rp->free_instances.first,
				 struct kretprobe_instance, uflist);
		ri->rp = rp;
		ri->task = current;
783 784 785 786 787 788

		if (rp->entry_handler && rp->entry_handler(ri, regs)) {
			spin_unlock_irqrestore(&kretprobe_lock, flags);
			return 0;
		}

789 790 791 792 793 794 795 796
		arch_prepare_kretprobe(ri, regs);

		/* XXX(hch): why is there no hlist_move_head? */
		hlist_del(&ri->uflist);
		hlist_add_head(&ri->uflist, &ri->rp->used_instances);
		hlist_add_head(&ri->hlist, kretprobe_inst_table_head(ri->task));
	} else
		rp->nmissed++;
797 798 799 800
	spin_unlock_irqrestore(&kretprobe_lock, flags);
	return 0;
}

801
int __kprobes register_kretprobe(struct kretprobe *rp)
802 803 804 805
{
	int ret = 0;
	struct kretprobe_instance *inst;
	int i;
806
	void *addr;
807 808

	if (kretprobe_blacklist_size) {
809 810 811
		addr = kprobe_addr(&rp->kp);
		if (!addr)
			return -EINVAL;
812 813 814 815 816 817

		for (i = 0; kretprobe_blacklist[i].name != NULL; i++) {
			if (kretprobe_blacklist[i].addr == addr)
				return -EINVAL;
		}
	}
818 819

	rp->kp.pre_handler = pre_handler_kretprobe;
820 821 822
	rp->kp.post_handler = NULL;
	rp->kp.fault_handler = NULL;
	rp->kp.break_handler = NULL;
823 824 825 826 827 828 829 830 831 832 833 834

	/* Pre-allocate memory for max kretprobe instances */
	if (rp->maxactive <= 0) {
#ifdef CONFIG_PREEMPT
		rp->maxactive = max(10, 2 * NR_CPUS);
#else
		rp->maxactive = NR_CPUS;
#endif
	}
	INIT_HLIST_HEAD(&rp->used_instances);
	INIT_HLIST_HEAD(&rp->free_instances);
	for (i = 0; i < rp->maxactive; i++) {
835 836
		inst = kmalloc(sizeof(struct kretprobe_instance) +
			       rp->data_size, GFP_KERNEL);
837 838 839 840 841 842 843 844 845 846
		if (inst == NULL) {
			free_rp_inst(rp);
			return -ENOMEM;
		}
		INIT_HLIST_NODE(&inst->uflist);
		hlist_add_head(&inst->uflist, &rp->free_instances);
	}

	rp->nmissed = 0;
	/* Establish function entry probe point */
847 848
	if ((ret = __register_kprobe(&rp->kp,
		(unsigned long)__builtin_return_address(0))) != 0)
849 850 851 852
		free_rp_inst(rp);
	return ret;
}

853
#else /* CONFIG_KRETPROBES */
854
int __kprobes register_kretprobe(struct kretprobe *rp)
855 856 857 858
{
	return -ENOSYS;
}

859 860 861 862 863
static int __kprobes pre_handler_kretprobe(struct kprobe *p,
					   struct pt_regs *regs)
{
	return 0;
}
864
#endif /* CONFIG_KRETPROBES */
865

866
void __kprobes unregister_kretprobe(struct kretprobe *rp)
867 868 869
{
	unsigned long flags;
	struct kretprobe_instance *ri;
870
	struct hlist_node *pos, *next;
871 872

	unregister_kprobe(&rp->kp);
873

874
	/* No race here */
875
	spin_lock_irqsave(&kretprobe_lock, flags);
876
	hlist_for_each_entry_safe(ri, pos, next, &rp->used_instances, uflist) {
877 878 879
		ri->rp = NULL;
		hlist_del(&ri->uflist);
	}
880
	spin_unlock_irqrestore(&kretprobe_lock, flags);
881
	free_rp_inst(rp);
882 883
}

L
Linus Torvalds 已提交
884 885 886
static int __init init_kprobes(void)
{
	int i, err = 0;
887 888 889 890 891
	unsigned long offset = 0, size = 0;
	char *modname, namebuf[128];
	const char *symbol_name;
	void *addr;
	struct kprobe_blackpoint *kb;
L
Linus Torvalds 已提交
892 893 894

	/* FIXME allocate the probe table, currently defined statically */
	/* initialize all list heads */
895
	for (i = 0; i < KPROBE_TABLE_SIZE; i++) {
L
Linus Torvalds 已提交
896
		INIT_HLIST_HEAD(&kprobe_table[i]);
897 898
		INIT_HLIST_HEAD(&kretprobe_inst_table[i]);
	}
L
Linus Torvalds 已提交
899

900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921
	/*
	 * Lookup and populate the kprobe_blacklist.
	 *
	 * Unlike the kretprobe blacklist, we'll need to determine
	 * the range of addresses that belong to the said functions,
	 * since a kprobe need not necessarily be at the beginning
	 * of a function.
	 */
	for (kb = kprobe_blacklist; kb->name != NULL; kb++) {
		kprobe_lookup_name(kb->name, addr);
		if (!addr)
			continue;

		kb->start_addr = (unsigned long)addr;
		symbol_name = kallsyms_lookup(kb->start_addr,
				&size, &offset, &modname, namebuf);
		if (!symbol_name)
			kb->range = 0;
		else
			kb->range = size;
	}

922 923 924 925 926 927 928 929 930 931 932
	if (kretprobe_blacklist_size) {
		/* lookup the function address from its name */
		for (i = 0; kretprobe_blacklist[i].name != NULL; i++) {
			kprobe_lookup_name(kretprobe_blacklist[i].name,
					   kretprobe_blacklist[i].addr);
			if (!kretprobe_blacklist[i].addr)
				printk("kretprobe: lookup failed: %s\n",
				       kretprobe_blacklist[i].name);
		}
	}

933 934 935
	/* By default, kprobes are enabled */
	kprobe_enabled = true;

936
	err = arch_init_kprobes();
937 938 939
	if (!err)
		err = register_die_notifier(&kprobe_exceptions_nb);

940 941
	if (!err)
		init_test_probes();
L
Linus Torvalds 已提交
942 943 944
	return err;
}

945 946
#ifdef CONFIG_DEBUG_FS
static void __kprobes report_probe(struct seq_file *pi, struct kprobe *p,
947
		const char *sym, int offset,char *modname)
948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988
{
	char *kprobe_type;

	if (p->pre_handler == pre_handler_kretprobe)
		kprobe_type = "r";
	else if (p->pre_handler == setjmp_pre_handler)
		kprobe_type = "j";
	else
		kprobe_type = "k";
	if (sym)
		seq_printf(pi, "%p  %s  %s+0x%x  %s\n", p->addr, kprobe_type,
			sym, offset, (modname ? modname : " "));
	else
		seq_printf(pi, "%p  %s  %p\n", p->addr, kprobe_type, p->addr);
}

static void __kprobes *kprobe_seq_start(struct seq_file *f, loff_t *pos)
{
	return (*pos < KPROBE_TABLE_SIZE) ? pos : NULL;
}

static void __kprobes *kprobe_seq_next(struct seq_file *f, void *v, loff_t *pos)
{
	(*pos)++;
	if (*pos >= KPROBE_TABLE_SIZE)
		return NULL;
	return pos;
}

static void __kprobes kprobe_seq_stop(struct seq_file *f, void *v)
{
	/* Nothing to do */
}

static int __kprobes show_kprobe_addr(struct seq_file *pi, void *v)
{
	struct hlist_head *head;
	struct hlist_node *node;
	struct kprobe *p, *kp;
	const char *sym = NULL;
	unsigned int i = *(loff_t *) v;
A
Alexey Dobriyan 已提交
989
	unsigned long offset = 0;
990 991 992 993 994
	char *modname, namebuf[128];

	head = &kprobe_table[i];
	preempt_disable();
	hlist_for_each_entry_rcu(p, node, head, hlist) {
A
Alexey Dobriyan 已提交
995
		sym = kallsyms_lookup((unsigned long)p->addr, NULL,
996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025
					&offset, &modname, namebuf);
		if (p->pre_handler == aggr_pre_handler) {
			list_for_each_entry_rcu(kp, &p->list, list)
				report_probe(pi, kp, sym, offset, modname);
		} else
			report_probe(pi, p, sym, offset, modname);
	}
	preempt_enable();
	return 0;
}

static struct seq_operations kprobes_seq_ops = {
	.start = kprobe_seq_start,
	.next  = kprobe_seq_next,
	.stop  = kprobe_seq_stop,
	.show  = show_kprobe_addr
};

static int __kprobes kprobes_open(struct inode *inode, struct file *filp)
{
	return seq_open(filp, &kprobes_seq_ops);
}

static struct file_operations debugfs_kprobes_operations = {
	.open           = kprobes_open,
	.read           = seq_read,
	.llseek         = seq_lseek,
	.release        = seq_release,
};

1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078
static void __kprobes enable_all_kprobes(void)
{
	struct hlist_head *head;
	struct hlist_node *node;
	struct kprobe *p;
	unsigned int i;

	mutex_lock(&kprobe_mutex);

	/* If kprobes are already enabled, just return */
	if (kprobe_enabled)
		goto already_enabled;

	for (i = 0; i < KPROBE_TABLE_SIZE; i++) {
		head = &kprobe_table[i];
		hlist_for_each_entry_rcu(p, node, head, hlist)
			arch_arm_kprobe(p);
	}

	kprobe_enabled = true;
	printk(KERN_INFO "Kprobes globally enabled\n");

already_enabled:
	mutex_unlock(&kprobe_mutex);
	return;
}

static void __kprobes disable_all_kprobes(void)
{
	struct hlist_head *head;
	struct hlist_node *node;
	struct kprobe *p;
	unsigned int i;

	mutex_lock(&kprobe_mutex);

	/* If kprobes are already disabled, just return */
	if (!kprobe_enabled)
		goto already_disabled;

	kprobe_enabled = false;
	printk(KERN_INFO "Kprobes globally disabled\n");
	for (i = 0; i < KPROBE_TABLE_SIZE; i++) {
		head = &kprobe_table[i];
		hlist_for_each_entry_rcu(p, node, head, hlist) {
			if (!arch_trampoline_kprobe(p))
				arch_disarm_kprobe(p);
		}
	}

	mutex_unlock(&kprobe_mutex);
	/* Allow all currently running kprobes to complete */
	synchronize_sched();
1079
	return;
1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135

already_disabled:
	mutex_unlock(&kprobe_mutex);
	return;
}

/*
 * XXX: The debugfs bool file interface doesn't allow for callbacks
 * when the bool state is switched. We can reuse that facility when
 * available
 */
static ssize_t read_enabled_file_bool(struct file *file,
	       char __user *user_buf, size_t count, loff_t *ppos)
{
	char buf[3];

	if (kprobe_enabled)
		buf[0] = '1';
	else
		buf[0] = '0';
	buf[1] = '\n';
	buf[2] = 0x00;
	return simple_read_from_buffer(user_buf, count, ppos, buf, 2);
}

static ssize_t write_enabled_file_bool(struct file *file,
	       const char __user *user_buf, size_t count, loff_t *ppos)
{
	char buf[32];
	int buf_size;

	buf_size = min(count, (sizeof(buf)-1));
	if (copy_from_user(buf, user_buf, buf_size))
		return -EFAULT;

	switch (buf[0]) {
	case 'y':
	case 'Y':
	case '1':
		enable_all_kprobes();
		break;
	case 'n':
	case 'N':
	case '0':
		disable_all_kprobes();
		break;
	}

	return count;
}

static struct file_operations fops_kp = {
	.read =         read_enabled_file_bool,
	.write =        write_enabled_file_bool,
};

1136 1137 1138
static int __kprobes debugfs_kprobe_init(void)
{
	struct dentry *dir, *file;
1139
	unsigned int value = 1;
1140 1141 1142 1143 1144

	dir = debugfs_create_dir("kprobes", NULL);
	if (!dir)
		return -ENOMEM;

R
Randy Dunlap 已提交
1145
	file = debugfs_create_file("list", 0444, dir, NULL,
1146 1147 1148 1149 1150 1151
				&debugfs_kprobes_operations);
	if (!file) {
		debugfs_remove(dir);
		return -ENOMEM;
	}

1152 1153 1154 1155 1156 1157 1158
	file = debugfs_create_file("enabled", 0600, dir,
					&value, &fops_kp);
	if (!file) {
		debugfs_remove(dir);
		return -ENOMEM;
	}

1159 1160 1161 1162 1163 1164 1165
	return 0;
}

late_initcall(debugfs_kprobe_init);
#endif /* CONFIG_DEBUG_FS */

module_init(init_kprobes);
L
Linus Torvalds 已提交
1166 1167 1168

EXPORT_SYMBOL_GPL(register_kprobe);
EXPORT_SYMBOL_GPL(unregister_kprobe);
1169 1170
EXPORT_SYMBOL_GPL(register_kprobes);
EXPORT_SYMBOL_GPL(unregister_kprobes);
L
Linus Torvalds 已提交
1171 1172
EXPORT_SYMBOL_GPL(register_jprobe);
EXPORT_SYMBOL_GPL(unregister_jprobe);
P
Peter Chubb 已提交
1173
#ifdef CONFIG_KPROBES
L
Linus Torvalds 已提交
1174
EXPORT_SYMBOL_GPL(jprobe_return);
P
Peter Chubb 已提交
1175 1176 1177
#endif

#ifdef CONFIG_KPROBES
1178 1179
EXPORT_SYMBOL_GPL(register_kretprobe);
EXPORT_SYMBOL_GPL(unregister_kretprobe);
P
Peter Chubb 已提交
1180
#endif