binder.c 151.1 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17
/* binder.c
 *
 * Android IPC Subsystem
 *
 * Copyright (C) 2007-2008 Google, Inc.
 *
 * This software is licensed under the terms of the GNU General Public
 * License version 2, as published by the Free Software Foundation, and
 * may be copied, distributed, and modified under those terms.
 *
 * This program is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 */

18 19 20 21 22 23 24 25 26 27 28 29 30 31 32
/*
 * Locking overview
 *
 * There are 3 main spinlocks which must be acquired in the
 * order shown:
 *
 * 1) proc->outer_lock : protects binder_ref
 *    binder_proc_lock() and binder_proc_unlock() are
 *    used to acq/rel.
 * 2) node->lock : protects most fields of binder_node.
 *    binder_node_lock() and binder_node_unlock() are
 *    used to acq/rel
 * 3) proc->inner_lock : protects the thread and node lists
 *    (proc->threads, proc->nodes) and all todo lists associated
 *    with the binder_proc (proc->todo, thread->todo,
33 34
 *    proc->delivered_death and node->async_todo), as well as
 *    thread->transaction_stack
35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51
 *    binder_inner_proc_lock() and binder_inner_proc_unlock()
 *    are used to acq/rel
 *
 * Any lock under procA must never be nested under any lock at the same
 * level or below on procB.
 *
 * Functions that require a lock held on entry indicate which lock
 * in the suffix of the function name:
 *
 * foo_olocked() : requires node->outer_lock
 * foo_nlocked() : requires node->lock
 * foo_ilocked() : requires proc->inner_lock
 * foo_oilocked(): requires proc->outer_lock and proc->inner_lock
 * foo_nilocked(): requires node->lock and proc->inner_lock
 * ...
 */

52 53
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt

54 55 56
#include <asm/cacheflush.h>
#include <linux/fdtable.h>
#include <linux/file.h>
57
#include <linux/freezer.h>
58 59 60 61 62 63 64
#include <linux/fs.h>
#include <linux/list.h>
#include <linux/miscdevice.h>
#include <linux/module.h>
#include <linux/mutex.h>
#include <linux/nsproxy.h>
#include <linux/poll.h>
65
#include <linux/debugfs.h>
66
#include <linux/rbtree.h>
67
#include <linux/sched/signal.h>
68
#include <linux/sched/mm.h>
69
#include <linux/seq_file.h>
70
#include <linux/uaccess.h>
71
#include <linux/pid_namespace.h>
72
#include <linux/security.h>
73
#include <linux/spinlock.h>
74

75 76 77 78 79
#ifdef CONFIG_ANDROID_BINDER_IPC_32BIT
#define BINDER_IPC_32BIT 1
#endif

#include <uapi/linux/android/binder.h>
80
#include "binder_alloc.h"
81
#include "binder_trace.h"
82

83
static HLIST_HEAD(binder_deferred_list);
84 85
static DEFINE_MUTEX(binder_deferred_lock);

86
static HLIST_HEAD(binder_devices);
87
static HLIST_HEAD(binder_procs);
88 89
static DEFINE_MUTEX(binder_procs_lock);

90
static HLIST_HEAD(binder_dead_nodes);
91
static DEFINE_SPINLOCK(binder_dead_nodes_lock);
92

93 94
static struct dentry *binder_debugfs_dir_entry_root;
static struct dentry *binder_debugfs_dir_entry_proc;
95
static atomic_t binder_last_id;
96

97 98 99
#define BINDER_DEBUG_ENTRY(name) \
static int binder_##name##_open(struct inode *inode, struct file *file) \
{ \
100
	return single_open(file, binder_##name##_show, inode->i_private); \
101 102 103 104 105 106 107 108 109 110 111 112
} \
\
static const struct file_operations binder_##name##_fops = { \
	.owner = THIS_MODULE, \
	.open = binder_##name##_open, \
	.read = seq_read, \
	.llseek = seq_lseek, \
	.release = single_release, \
}

static int binder_proc_show(struct seq_file *m, void *unused);
BINDER_DEBUG_ENTRY(proc);
113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138

/* This is only defined in include/asm-arm/sizes.h */
#ifndef SZ_1K
#define SZ_1K                               0x400
#endif

#ifndef SZ_4M
#define SZ_4M                               0x400000
#endif

#define FORBIDDEN_MMAP_FLAGS                (VM_WRITE)

enum {
	BINDER_DEBUG_USER_ERROR             = 1U << 0,
	BINDER_DEBUG_FAILED_TRANSACTION     = 1U << 1,
	BINDER_DEBUG_DEAD_TRANSACTION       = 1U << 2,
	BINDER_DEBUG_OPEN_CLOSE             = 1U << 3,
	BINDER_DEBUG_DEAD_BINDER            = 1U << 4,
	BINDER_DEBUG_DEATH_NOTIFICATION     = 1U << 5,
	BINDER_DEBUG_READ_WRITE             = 1U << 6,
	BINDER_DEBUG_USER_REFS              = 1U << 7,
	BINDER_DEBUG_THREADS                = 1U << 8,
	BINDER_DEBUG_TRANSACTION            = 1U << 9,
	BINDER_DEBUG_TRANSACTION_COMPLETE   = 1U << 10,
	BINDER_DEBUG_FREE_BUFFER            = 1U << 11,
	BINDER_DEBUG_INTERNAL_REFS          = 1U << 12,
139
	BINDER_DEBUG_PRIORITY_CAP           = 1U << 13,
140
	BINDER_DEBUG_SPINLOCKS              = 1U << 14,
141 142 143 144 145
};
static uint32_t binder_debug_mask = BINDER_DEBUG_USER_ERROR |
	BINDER_DEBUG_FAILED_TRANSACTION | BINDER_DEBUG_DEAD_TRANSACTION;
module_param_named(debug_mask, binder_debug_mask, uint, S_IWUSR | S_IRUGO);

146 147 148
static char *binder_devices_param = CONFIG_ANDROID_BINDER_DEVICES;
module_param_named(devices, binder_devices_param, charp, 0444);

149 150 151 152 153 154 155
static DECLARE_WAIT_QUEUE_HEAD(binder_user_error_wait);
static int binder_stop_on_user_error;

static int binder_set_stop_on_user_error(const char *val,
					 struct kernel_param *kp)
{
	int ret;
156

157 158 159 160 161 162 163 164 165 166 167
	ret = param_set_int(val, kp);
	if (binder_stop_on_user_error < 2)
		wake_up(&binder_user_error_wait);
	return ret;
}
module_param_call(stop_on_user_error, binder_set_stop_on_user_error,
	param_get_int, &binder_stop_on_user_error, S_IWUSR | S_IRUGO);

#define binder_debug(mask, x...) \
	do { \
		if (binder_debug_mask & mask) \
168
			pr_info(x); \
169 170 171 172 173
	} while (0)

#define binder_user_error(x...) \
	do { \
		if (binder_debug_mask & BINDER_DEBUG_USER_ERROR) \
174
			pr_info(x); \
175 176 177 178
		if (binder_stop_on_user_error) \
			binder_stop_on_user_error = 2; \
	} while (0)

179 180 181 182 183
#define to_flat_binder_object(hdr) \
	container_of(hdr, struct flat_binder_object, hdr)

#define to_binder_fd_object(hdr) container_of(hdr, struct binder_fd_object, hdr)

184 185 186
#define to_binder_buffer_object(hdr) \
	container_of(hdr, struct binder_buffer_object, hdr)

187 188 189
#define to_binder_fd_array_object(hdr) \
	container_of(hdr, struct binder_fd_array_object, hdr)

190 191 192 193 194 195 196 197 198 199 200 201
enum binder_stat_types {
	BINDER_STAT_PROC,
	BINDER_STAT_THREAD,
	BINDER_STAT_NODE,
	BINDER_STAT_REF,
	BINDER_STAT_DEATH,
	BINDER_STAT_TRANSACTION,
	BINDER_STAT_TRANSACTION_COMPLETE,
	BINDER_STAT_COUNT
};

struct binder_stats {
202 203 204 205
	atomic_t br[_IOC_NR(BR_FAILED_REPLY) + 1];
	atomic_t bc[_IOC_NR(BC_REPLY_SG) + 1];
	atomic_t obj_created[BINDER_STAT_COUNT];
	atomic_t obj_deleted[BINDER_STAT_COUNT];
206 207 208 209 210 211
};

static struct binder_stats binder_stats;

static inline void binder_stats_deleted(enum binder_stat_types type)
{
212
	atomic_inc(&binder_stats.obj_deleted[type]);
213 214 215 216
}

static inline void binder_stats_created(enum binder_stat_types type)
{
217
	atomic_inc(&binder_stats.obj_created[type]);
218 219 220 221
}

struct binder_transaction_log_entry {
	int debug_id;
222
	int debug_id_done;
223 224 225 226 227 228 229 230 231
	int call_type;
	int from_proc;
	int from_thread;
	int target_handle;
	int to_proc;
	int to_thread;
	int to_node;
	int data_size;
	int offsets_size;
232 233 234
	int return_error_line;
	uint32_t return_error;
	uint32_t return_error_param;
235
	const char *context_name;
236 237
};
struct binder_transaction_log {
238 239
	atomic_t cur;
	bool full;
240 241 242 243 244 245 246 247 248
	struct binder_transaction_log_entry entry[32];
};
static struct binder_transaction_log binder_transaction_log;
static struct binder_transaction_log binder_transaction_log_failed;

static struct binder_transaction_log_entry *binder_transaction_log_add(
	struct binder_transaction_log *log)
{
	struct binder_transaction_log_entry *e;
249
	unsigned int cur = atomic_inc_return(&log->cur);
250

251
	if (cur >= ARRAY_SIZE(log->entry))
252
		log->full = 1;
253 254 255 256 257 258 259 260 261
	e = &log->entry[cur % ARRAY_SIZE(log->entry)];
	WRITE_ONCE(e->debug_id_done, 0);
	/*
	 * write-barrier to synchronize access to e->debug_id_done.
	 * We make sure the initialized 0 value is seen before
	 * memset() other fields are zeroed by memset.
	 */
	smp_wmb();
	memset(e, 0, sizeof(*e));
262 263 264
	return e;
}

265 266
struct binder_context {
	struct binder_node *binder_context_mgr_node;
267 268
	struct mutex context_mgr_node_lock;

269
	kuid_t binder_context_mgr_uid;
270
	const char *name;
271 272
};

273 274 275 276
struct binder_device {
	struct hlist_node hlist;
	struct miscdevice miscdev;
	struct binder_context context;
277 278
};

279 280 281 282 283 284 285
/**
 * struct binder_work - work enqueued on a worklist
 * @entry:             node enqueued on list
 * @type:              type of work to be performed
 *
 * There are separate work lists for proc, thread, and node (async).
 */
286 287
struct binder_work {
	struct list_head entry;
288

289 290 291
	enum {
		BINDER_WORK_TRANSACTION = 1,
		BINDER_WORK_TRANSACTION_COMPLETE,
292
		BINDER_WORK_RETURN_ERROR,
293 294 295 296 297 298 299
		BINDER_WORK_NODE,
		BINDER_WORK_DEAD_BINDER,
		BINDER_WORK_DEAD_BINDER_AND_CLEAR,
		BINDER_WORK_CLEAR_DEATH_NOTIFICATION,
	} type;
};

300 301 302 303 304
struct binder_error {
	struct binder_work work;
	uint32_t cmd;
};

305 306 307 308 309 310
/**
 * struct binder_node - binder node bookkeeping
 * @debug_id:             unique ID for debugging
 *                        (invariant after initialized)
 * @lock:                 lock for node fields
 * @work:                 worklist element for node work
311
 *                        (protected by @proc->inner_lock)
312
 * @rb_node:              element for proc->nodes tree
313
 *                        (protected by @proc->inner_lock)
314 315 316 317 318
 * @dead_node:            element for binder_dead_nodes list
 *                        (protected by binder_dead_nodes_lock)
 * @proc:                 binder_proc that owns this node
 *                        (invariant after initialized)
 * @refs:                 list of references on this node
319
 *                        (protected by @lock)
320 321
 * @internal_strong_refs: used to take strong references when
 *                        initiating a transaction
322 323
 *                        (protected by @proc->inner_lock if @proc
 *                        and by @lock)
324
 * @local_weak_refs:      weak user refs from local process
325 326
 *                        (protected by @proc->inner_lock if @proc
 *                        and by @lock)
327
 * @local_strong_refs:    strong user refs from local process
328 329
 *                        (protected by @proc->inner_lock if @proc
 *                        and by @lock)
330
 * @tmp_refs:             temporary kernel refs
331 332 333 334 335
 *                        (protected by @proc->inner_lock while @proc
 *                        is valid, and by binder_dead_nodes_lock
 *                        if @proc is NULL. During inc/dec and node release
 *                        it is also protected by @lock to provide safety
 *                        as the node dies and @proc becomes NULL)
336 337 338 339 340
 * @ptr:                  userspace pointer for node
 *                        (invariant, no lock needed)
 * @cookie:               userspace cookie for node
 *                        (invariant, no lock needed)
 * @has_strong_ref:       userspace notified of strong ref
341 342
 *                        (protected by @proc->inner_lock if @proc
 *                        and by @lock)
343
 * @pending_strong_ref:   userspace has acked notification of strong ref
344 345
 *                        (protected by @proc->inner_lock if @proc
 *                        and by @lock)
346
 * @has_weak_ref:         userspace notified of weak ref
347 348
 *                        (protected by @proc->inner_lock if @proc
 *                        and by @lock)
349
 * @pending_weak_ref:     userspace has acked notification of weak ref
350 351
 *                        (protected by @proc->inner_lock if @proc
 *                        and by @lock)
352
 * @has_async_transaction: async transaction to node in progress
353
 *                        (protected by @lock)
354 355 356 357 358
 * @accept_fds:           file descriptor operations supported for node
 *                        (invariant after initialized)
 * @min_priority:         minimum scheduling priority
 *                        (invariant after initialized)
 * @async_todo:           list of async work items
359
 *                        (protected by @proc->inner_lock)
360 361 362
 *
 * Bookkeeping structure for binder nodes.
 */
363 364
struct binder_node {
	int debug_id;
365
	spinlock_t lock;
366 367 368 369 370 371 372 373 374 375
	struct binder_work work;
	union {
		struct rb_node rb_node;
		struct hlist_node dead_node;
	};
	struct binder_proc *proc;
	struct hlist_head refs;
	int internal_strong_refs;
	int local_weak_refs;
	int local_strong_refs;
376
	int tmp_refs;
377 378
	binder_uintptr_t ptr;
	binder_uintptr_t cookie;
379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396
	struct {
		/*
		 * bitfield elements protected by
		 * proc inner_lock
		 */
		u8 has_strong_ref:1;
		u8 pending_strong_ref:1;
		u8 has_weak_ref:1;
		u8 pending_weak_ref:1;
	};
	struct {
		/*
		 * invariant after initialization
		 */
		u8 accept_fds:1;
		u8 min_priority;
	};
	bool has_async_transaction;
397 398 399 400
	struct list_head async_todo;
};

struct binder_ref_death {
401 402 403 404 405
	/**
	 * @work: worklist element for death notifications
	 *        (protected by inner_lock of the proc that
	 *        this ref belongs to)
	 */
406
	struct binder_work work;
407
	binder_uintptr_t cookie;
408 409
};

410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434
/**
 * struct binder_ref_data - binder_ref counts and id
 * @debug_id:        unique ID for the ref
 * @desc:            unique userspace handle for ref
 * @strong:          strong ref count (debugging only if not locked)
 * @weak:            weak ref count (debugging only if not locked)
 *
 * Structure to hold ref count and ref id information. Since
 * the actual ref can only be accessed with a lock, this structure
 * is used to return information about the ref to callers of
 * ref inc/dec functions.
 */
struct binder_ref_data {
	int debug_id;
	uint32_t desc;
	int strong;
	int weak;
};

/**
 * struct binder_ref - struct to track references on nodes
 * @data:        binder_ref_data containing id, handle, and current refcounts
 * @rb_node_desc: node for lookup by @data.desc in proc's rb_tree
 * @rb_node_node: node for lookup by @node in proc's rb_tree
 * @node_entry:  list entry for node->refs list in target node
435
 *               (protected by @node->lock)
436 437 438 439 440
 * @proc:        binder_proc containing ref
 * @node:        binder_node of target node. When cleaning up a
 *               ref for deletion in binder_cleanup_ref, a non-NULL
 *               @node indicates the node must be freed
 * @death:       pointer to death notification (ref_death) if requested
441
 *               (protected by @node->lock)
442 443 444 445
 *
 * Structure to track references from procA to target node (on procB). This
 * structure is unsafe to access without holding @proc->outer_lock.
 */
446 447 448 449 450
struct binder_ref {
	/* Lookups needed: */
	/*   node + proc => ref (transaction) */
	/*   desc + proc => ref (transaction, inc/dec ref) */
	/*   node => refs + procs (proc exit) */
451
	struct binder_ref_data data;
452 453 454 455 456 457 458 459 460 461 462 463 464 465
	struct rb_node rb_node_desc;
	struct rb_node rb_node_node;
	struct hlist_node node_entry;
	struct binder_proc *proc;
	struct binder_node *node;
	struct binder_ref_death *death;
};

enum binder_deferred_state {
	BINDER_DEFERRED_PUT_FILES    = 0x01,
	BINDER_DEFERRED_FLUSH        = 0x02,
	BINDER_DEFERRED_RELEASE      = 0x04,
};

466 467 468 469
/**
 * struct binder_proc - binder process bookkeeping
 * @proc_node:            element for binder_procs list
 * @threads:              rbtree of binder_threads in this proc
470
 *                        (protected by @inner_lock)
471 472
 * @nodes:                rbtree of binder nodes associated with
 *                        this proc ordered by node->ptr
473
 *                        (protected by @inner_lock)
474
 * @refs_by_desc:         rbtree of refs ordered by ref->desc
475
 *                        (protected by @outer_lock)
476
 * @refs_by_node:         rbtree of refs ordered by ref->node
477
 *                        (protected by @outer_lock)
478 479 480 481 482 483 484 485 486 487 488 489
 * @pid                   PID of group_leader of process
 *                        (invariant after initialized)
 * @tsk                   task_struct for group_leader of process
 *                        (invariant after initialized)
 * @files                 files_struct for process
 *                        (invariant after initialized)
 * @deferred_work_node:   element for binder_deferred_list
 *                        (protected by binder_deferred_lock)
 * @deferred_work:        bitmap of deferred work to perform
 *                        (protected by binder_deferred_lock)
 * @is_dead:              process is dead and awaiting free
 *                        when outstanding transactions are cleaned up
490
 *                        (protected by @inner_lock)
491
 * @todo:                 list of work for this process
492
 *                        (protected by @inner_lock)
493 494 495 496 497
 * @wait:                 wait queue head to wait for proc work
 *                        (invariant after initialized)
 * @stats:                per-process binder statistics
 *                        (atomics, no lock needed)
 * @delivered_death:      list of delivered death notification
498
 *                        (protected by @inner_lock)
499
 * @max_threads:          cap on number of binder threads
500
 *                        (protected by @inner_lock)
501 502 503
 * @requested_threads:    number of binder threads requested but not
 *                        yet started. In current implementation, can
 *                        only be 0 or 1.
504
 *                        (protected by @inner_lock)
505
 * @requested_threads_started: number binder threads started
506
 *                        (protected by @inner_lock)
507
 * @ready_threads:        number of threads waiting for proc work
508
 *                        (protected by @inner_lock)
509
 * @tmp_ref:              temporary reference to indicate proc is in use
510
 *                        (protected by @inner_lock)
511 512 513 514 515 516 517 518 519 520 521 522
 * @default_priority:     default scheduler priority
 *                        (invariant after initialized)
 * @debugfs_entry:        debugfs node
 * @alloc:                binder allocator bookkeeping
 * @context:              binder_context for this proc
 *                        (invariant after initialized)
 * @inner_lock:           can nest under outer_lock and/or node lock
 * @outer_lock:           no nesting under innor or node lock
 *                        Lock order: 1) outer, 2) node, 3) inner
 *
 * Bookkeeping structure for binder processes
 */
523 524 525 526 527 528 529 530 531 532 533
struct binder_proc {
	struct hlist_node proc_node;
	struct rb_root threads;
	struct rb_root nodes;
	struct rb_root refs_by_desc;
	struct rb_root refs_by_node;
	int pid;
	struct task_struct *tsk;
	struct files_struct *files;
	struct hlist_node deferred_work_node;
	int deferred_work;
534
	bool is_dead;
535 536 537 538 539 540 541 542 543

	struct list_head todo;
	wait_queue_head_t wait;
	struct binder_stats stats;
	struct list_head delivered_death;
	int max_threads;
	int requested_threads;
	int requested_threads_started;
	int ready_threads;
544
	int tmp_ref;
545
	long default_priority;
546
	struct dentry *debugfs_entry;
547
	struct binder_alloc alloc;
548
	struct binder_context *context;
549 550
	spinlock_t inner_lock;
	spinlock_t outer_lock;
551 552 553 554 555 556 557 558 559 560
};

enum {
	BINDER_LOOPER_STATE_REGISTERED  = 0x01,
	BINDER_LOOPER_STATE_ENTERED     = 0x02,
	BINDER_LOOPER_STATE_EXITED      = 0x04,
	BINDER_LOOPER_STATE_INVALID     = 0x08,
	BINDER_LOOPER_STATE_WAITING     = 0x10,
};

561 562 563 564 565
/**
 * struct binder_thread - binder thread bookkeeping
 * @proc:                 binder process for this thread
 *                        (invariant after initialization)
 * @rb_node:              element for proc->threads rbtree
566
 *                        (protected by @proc->inner_lock)
567 568 569 570 571 572 573
 * @pid:                  PID for this thread
 *                        (invariant after initialization)
 * @looper:               bitmap of looping state
 *                        (only accessed by this thread)
 * @looper_needs_return:  looping thread needs to exit driver
 *                        (no lock needed)
 * @transaction_stack:    stack of in-progress transactions for this thread
574
 *                        (protected by @proc->inner_lock)
575
 * @todo:                 list of work to do for this thread
576
 *                        (protected by @proc->inner_lock)
577 578 579
 * @return_error:         transaction errors reported by this thread
 *                        (only accessed by this thread)
 * @reply_error:          transaction errors reported by target thread
580
 *                        (protected by @proc->inner_lock)
581 582 583 584 585 586 587 588
 * @wait:                 wait queue for thread work
 * @stats:                per-thread statistics
 *                        (atomics, no lock needed)
 * @tmp_ref:              temporary reference to indicate thread is in use
 *                        (atomic since @proc->inner_lock cannot
 *                        always be acquired)
 * @is_dead:              thread is dead and awaiting free
 *                        when outstanding transactions are cleaned up
589
 *                        (protected by @proc->inner_lock)
590 591 592
 *
 * Bookkeeping structure for binder threads.
 */
593 594 595 596
struct binder_thread {
	struct binder_proc *proc;
	struct rb_node rb_node;
	int pid;
597 598
	int looper;              /* only modified by this thread */
	bool looper_need_return; /* can be written by other thread */
599 600
	struct binder_transaction *transaction_stack;
	struct list_head todo;
601 602
	struct binder_error return_error;
	struct binder_error reply_error;
603 604
	wait_queue_head_t wait;
	struct binder_stats stats;
605 606
	atomic_t tmp_ref;
	bool is_dead;
607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624
};

struct binder_transaction {
	int debug_id;
	struct binder_work work;
	struct binder_thread *from;
	struct binder_transaction *from_parent;
	struct binder_proc *to_proc;
	struct binder_thread *to_thread;
	struct binder_transaction *to_parent;
	unsigned need_reply:1;
	/* unsigned is_dead:1; */	/* not used at the moment */

	struct binder_buffer *buffer;
	unsigned int	code;
	unsigned int	flags;
	long	priority;
	long	saved_priority;
625
	kuid_t	sender_euid;
626 627 628 629 630 631 632
	/**
	 * @lock:  protects @from, @to_proc, and @to_thread
	 *
	 * @from, @to_proc, and @to_thread can be set to NULL
	 * during thread teardown
	 */
	spinlock_t lock;
633 634
};

635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725
/**
 * binder_proc_lock() - Acquire outer lock for given binder_proc
 * @proc:         struct binder_proc to acquire
 *
 * Acquires proc->outer_lock. Used to protect binder_ref
 * structures associated with the given proc.
 */
#define binder_proc_lock(proc) _binder_proc_lock(proc, __LINE__)
static void
_binder_proc_lock(struct binder_proc *proc, int line)
{
	binder_debug(BINDER_DEBUG_SPINLOCKS,
		     "%s: line=%d\n", __func__, line);
	spin_lock(&proc->outer_lock);
}

/**
 * binder_proc_unlock() - Release spinlock for given binder_proc
 * @proc:         struct binder_proc to acquire
 *
 * Release lock acquired via binder_proc_lock()
 */
#define binder_proc_unlock(_proc) _binder_proc_unlock(_proc, __LINE__)
static void
_binder_proc_unlock(struct binder_proc *proc, int line)
{
	binder_debug(BINDER_DEBUG_SPINLOCKS,
		     "%s: line=%d\n", __func__, line);
	spin_unlock(&proc->outer_lock);
}

/**
 * binder_inner_proc_lock() - Acquire inner lock for given binder_proc
 * @proc:         struct binder_proc to acquire
 *
 * Acquires proc->inner_lock. Used to protect todo lists
 */
#define binder_inner_proc_lock(proc) _binder_inner_proc_lock(proc, __LINE__)
static void
_binder_inner_proc_lock(struct binder_proc *proc, int line)
{
	binder_debug(BINDER_DEBUG_SPINLOCKS,
		     "%s: line=%d\n", __func__, line);
	spin_lock(&proc->inner_lock);
}

/**
 * binder_inner_proc_unlock() - Release inner lock for given binder_proc
 * @proc:         struct binder_proc to acquire
 *
 * Release lock acquired via binder_inner_proc_lock()
 */
#define binder_inner_proc_unlock(proc) _binder_inner_proc_unlock(proc, __LINE__)
static void
_binder_inner_proc_unlock(struct binder_proc *proc, int line)
{
	binder_debug(BINDER_DEBUG_SPINLOCKS,
		     "%s: line=%d\n", __func__, line);
	spin_unlock(&proc->inner_lock);
}

/**
 * binder_node_lock() - Acquire spinlock for given binder_node
 * @node:         struct binder_node to acquire
 *
 * Acquires node->lock. Used to protect binder_node fields
 */
#define binder_node_lock(node) _binder_node_lock(node, __LINE__)
static void
_binder_node_lock(struct binder_node *node, int line)
{
	binder_debug(BINDER_DEBUG_SPINLOCKS,
		     "%s: line=%d\n", __func__, line);
	spin_lock(&node->lock);
}

/**
 * binder_node_unlock() - Release spinlock for given binder_proc
 * @node:         struct binder_node to acquire
 *
 * Release lock acquired via binder_node_lock()
 */
#define binder_node_unlock(node) _binder_node_unlock(node, __LINE__)
static void
_binder_node_unlock(struct binder_node *node, int line)
{
	binder_debug(BINDER_DEBUG_SPINLOCKS,
		     "%s: line=%d\n", __func__, line);
	spin_unlock(&node->lock);
}

726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762
/**
 * binder_node_inner_lock() - Acquire node and inner locks
 * @node:         struct binder_node to acquire
 *
 * Acquires node->lock. If node->proc also acquires
 * proc->inner_lock. Used to protect binder_node fields
 */
#define binder_node_inner_lock(node) _binder_node_inner_lock(node, __LINE__)
static void
_binder_node_inner_lock(struct binder_node *node, int line)
{
	binder_debug(BINDER_DEBUG_SPINLOCKS,
		     "%s: line=%d\n", __func__, line);
	spin_lock(&node->lock);
	if (node->proc)
		binder_inner_proc_lock(node->proc);
}

/**
 * binder_node_unlock() - Release node and inner locks
 * @node:         struct binder_node to acquire
 *
 * Release lock acquired via binder_node_lock()
 */
#define binder_node_inner_unlock(node) _binder_node_inner_unlock(node, __LINE__)
static void
_binder_node_inner_unlock(struct binder_node *node, int line)
{
	struct binder_proc *proc = node->proc;

	binder_debug(BINDER_DEBUG_SPINLOCKS,
		     "%s: line=%d\n", __func__, line);
	if (proc)
		binder_inner_proc_unlock(proc);
	spin_unlock(&node->lock);
}

763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867
static bool binder_worklist_empty_ilocked(struct list_head *list)
{
	return list_empty(list);
}

/**
 * binder_worklist_empty() - Check if no items on the work list
 * @proc:       binder_proc associated with list
 * @list:	list to check
 *
 * Return: true if there are no items on list, else false
 */
static bool binder_worklist_empty(struct binder_proc *proc,
				  struct list_head *list)
{
	bool ret;

	binder_inner_proc_lock(proc);
	ret = binder_worklist_empty_ilocked(list);
	binder_inner_proc_unlock(proc);
	return ret;
}

static void
binder_enqueue_work_ilocked(struct binder_work *work,
			   struct list_head *target_list)
{
	BUG_ON(target_list == NULL);
	BUG_ON(work->entry.next && !list_empty(&work->entry));
	list_add_tail(&work->entry, target_list);
}

/**
 * binder_enqueue_work() - Add an item to the work list
 * @proc:         binder_proc associated with list
 * @work:         struct binder_work to add to list
 * @target_list:  list to add work to
 *
 * Adds the work to the specified list. Asserts that work
 * is not already on a list.
 */
static void
binder_enqueue_work(struct binder_proc *proc,
		    struct binder_work *work,
		    struct list_head *target_list)
{
	binder_inner_proc_lock(proc);
	binder_enqueue_work_ilocked(work, target_list);
	binder_inner_proc_unlock(proc);
}

static void
binder_dequeue_work_ilocked(struct binder_work *work)
{
	list_del_init(&work->entry);
}

/**
 * binder_dequeue_work() - Removes an item from the work list
 * @proc:         binder_proc associated with list
 * @work:         struct binder_work to remove from list
 *
 * Removes the specified work item from whatever list it is on.
 * Can safely be called if work is not on any list.
 */
static void
binder_dequeue_work(struct binder_proc *proc, struct binder_work *work)
{
	binder_inner_proc_lock(proc);
	binder_dequeue_work_ilocked(work);
	binder_inner_proc_unlock(proc);
}

static struct binder_work *binder_dequeue_work_head_ilocked(
					struct list_head *list)
{
	struct binder_work *w;

	w = list_first_entry_or_null(list, struct binder_work, entry);
	if (w)
		list_del_init(&w->entry);
	return w;
}

/**
 * binder_dequeue_work_head() - Dequeues the item at head of list
 * @proc:         binder_proc associated with list
 * @list:         list to dequeue head
 *
 * Removes the head of the list if there are items on the list
 *
 * Return: pointer dequeued binder_work, NULL if list was empty
 */
static struct binder_work *binder_dequeue_work_head(
					struct binder_proc *proc,
					struct list_head *list)
{
	struct binder_work *w;

	binder_inner_proc_lock(proc);
	w = binder_dequeue_work_head_ilocked(list);
	binder_inner_proc_unlock(proc);
	return w;
}

868 869
static void
binder_defer_work(struct binder_proc *proc, enum binder_deferred_state defer);
870 871
static void binder_free_thread(struct binder_thread *thread);
static void binder_free_proc(struct binder_proc *proc);
872
static void binder_inc_node_tmpref_ilocked(struct binder_node *node);
873

874
static int task_get_unused_fd_flags(struct binder_proc *proc, int flags)
875 876 877 878 879 880 881 882
{
	struct files_struct *files = proc->files;
	unsigned long rlim_cur;
	unsigned long irqs;

	if (files == NULL)
		return -ESRCH;

A
Al Viro 已提交
883 884
	if (!lock_task_sighand(proc->tsk, &irqs))
		return -EMFILE;
885

A
Al Viro 已提交
886 887
	rlim_cur = task_rlimit(proc->tsk, RLIMIT_NOFILE);
	unlock_task_sighand(proc->tsk, &irqs);
888

A
Al Viro 已提交
889
	return __alloc_fd(files, 0, rlim_cur, flags);
890 891 892 893 894 895 896 897
}

/*
 * copied from fd_install
 */
static void task_fd_install(
	struct binder_proc *proc, unsigned int fd, struct file *file)
{
898 899
	if (proc->files)
		__fd_install(proc->files, fd, file);
900 901 902 903 904 905 906 907 908
}

/*
 * copied from sys_close
 */
static long task_close_fd(struct binder_proc *proc, unsigned int fd)
{
	int retval;

909
	if (proc->files == NULL)
910 911
		return -ESRCH;

912
	retval = __close_fd(proc->files, fd);
913 914 915 916 917 918 919 920 921 922 923 924 925
	/* can't restart close syscall because file table entry was cleared */
	if (unlikely(retval == -ERESTARTSYS ||
		     retval == -ERESTARTNOINTR ||
		     retval == -ERESTARTNOHAND ||
		     retval == -ERESTART_RESTARTBLOCK))
		retval = -EINTR;

	return retval;
}

static void binder_set_nice(long nice)
{
	long min_nice;
926

927 928 929 930
	if (can_nice(current, nice)) {
		set_user_nice(current, nice);
		return;
	}
931
	min_nice = rlimit_to_nice(rlimit(RLIMIT_NICE));
932
	binder_debug(BINDER_DEBUG_PRIORITY_CAP,
933 934
		     "%d: nice value %ld not allowed use %ld instead\n",
		      current->pid, nice, min_nice);
935
	set_user_nice(current, min_nice);
936
	if (min_nice <= MAX_NICE)
937
		return;
938
	binder_user_error("%d RLIMIT_NICE not set\n", current->pid);
939 940
}

941 942
static struct binder_node *binder_get_node_ilocked(struct binder_proc *proc,
						   binder_uintptr_t ptr)
943 944 945 946
{
	struct rb_node *n = proc->nodes.rb_node;
	struct binder_node *node;

947 948
	BUG_ON(!spin_is_locked(&proc->inner_lock));

949 950 951 952 953 954 955
	while (n) {
		node = rb_entry(n, struct binder_node, rb_node);

		if (ptr < node->ptr)
			n = n->rb_left;
		else if (ptr > node->ptr)
			n = n->rb_right;
956 957 958 959 960 961
		else {
			/*
			 * take an implicit weak reference
			 * to ensure node stays alive until
			 * call to binder_put_node()
			 */
962
			binder_inc_node_tmpref_ilocked(node);
963
			return node;
964
		}
965 966 967 968
	}
	return NULL;
}

969 970 971 972 973 974 975 976 977 978 979 980 981 982 983
static struct binder_node *binder_get_node(struct binder_proc *proc,
					   binder_uintptr_t ptr)
{
	struct binder_node *node;

	binder_inner_proc_lock(proc);
	node = binder_get_node_ilocked(proc, ptr);
	binder_inner_proc_unlock(proc);
	return node;
}

static struct binder_node *binder_init_node_ilocked(
						struct binder_proc *proc,
						struct binder_node *new_node,
						struct flat_binder_object *fp)
984 985 986 987
{
	struct rb_node **p = &proc->nodes.rb_node;
	struct rb_node *parent = NULL;
	struct binder_node *node;
988 989 990
	binder_uintptr_t ptr = fp ? fp->binder : 0;
	binder_uintptr_t cookie = fp ? fp->cookie : 0;
	__u32 flags = fp ? fp->flags : 0;
991

992
	BUG_ON(!spin_is_locked(&proc->inner_lock));
993
	while (*p) {
994

995 996 997 998 999 1000 1001
		parent = *p;
		node = rb_entry(parent, struct binder_node, rb_node);

		if (ptr < node->ptr)
			p = &(*p)->rb_left;
		else if (ptr > node->ptr)
			p = &(*p)->rb_right;
1002 1003 1004 1005 1006 1007 1008 1009 1010
		else {
			/*
			 * A matching node is already in
			 * the rb tree. Abandon the init
			 * and return it.
			 */
			binder_inc_node_tmpref_ilocked(node);
			return node;
		}
1011
	}
1012
	node = new_node;
1013
	binder_stats_created(BINDER_STAT_NODE);
1014
	node->tmp_refs++;
1015 1016
	rb_link_node(&node->rb_node, parent, p);
	rb_insert_color(&node->rb_node, &proc->nodes);
1017
	node->debug_id = atomic_inc_return(&binder_last_id);
1018 1019 1020 1021
	node->proc = proc;
	node->ptr = ptr;
	node->cookie = cookie;
	node->work.type = BINDER_WORK_NODE;
1022 1023
	node->min_priority = flags & FLAT_BINDER_FLAG_PRIORITY_MASK;
	node->accept_fds = !!(flags & FLAT_BINDER_FLAG_ACCEPTS_FDS);
1024
	spin_lock_init(&node->lock);
1025 1026 1027
	INIT_LIST_HEAD(&node->work.entry);
	INIT_LIST_HEAD(&node->async_todo);
	binder_debug(BINDER_DEBUG_INTERNAL_REFS,
1028
		     "%d:%d node %d u%016llx c%016llx created\n",
1029
		     proc->pid, current->pid, node->debug_id,
1030
		     (u64)node->ptr, (u64)node->cookie);
1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051

	return node;
}

static struct binder_node *binder_new_node(struct binder_proc *proc,
					   struct flat_binder_object *fp)
{
	struct binder_node *node;
	struct binder_node *new_node = kzalloc(sizeof(*node), GFP_KERNEL);

	if (!new_node)
		return NULL;
	binder_inner_proc_lock(proc);
	node = binder_init_node_ilocked(proc, new_node, fp);
	binder_inner_proc_unlock(proc);
	if (node != new_node)
		/*
		 * The node was already added by another thread
		 */
		kfree(new_node);

1052 1053 1054
	return node;
}

1055
static void binder_free_node(struct binder_node *node)
1056
{
1057 1058 1059 1060
	kfree(node);
	binder_stats_deleted(BINDER_STAT_NODE);
}

1061 1062 1063
static int binder_inc_node_nilocked(struct binder_node *node, int strong,
				    int internal,
				    struct list_head *target_list)
1064
{
1065 1066 1067 1068 1069
	struct binder_proc *proc = node->proc;

	BUG_ON(!spin_is_locked(&node->lock));
	if (proc)
		BUG_ON(!spin_is_locked(&proc->inner_lock));
1070 1071 1072 1073
	if (strong) {
		if (internal) {
			if (target_list == NULL &&
			    node->internal_strong_refs == 0 &&
1074 1075 1076
			    !(node->proc &&
			      node == node->proc->context->binder_context_mgr_node &&
			      node->has_strong_ref)) {
1077 1078
				pr_err("invalid inc strong node for %d\n",
					node->debug_id);
1079 1080 1081 1082 1083 1084
				return -EINVAL;
			}
			node->internal_strong_refs++;
		} else
			node->local_strong_refs++;
		if (!node->has_strong_ref && target_list) {
1085 1086
			binder_dequeue_work_ilocked(&node->work);
			binder_enqueue_work_ilocked(&node->work, target_list);
1087 1088 1089 1090 1091 1092
		}
	} else {
		if (!internal)
			node->local_weak_refs++;
		if (!node->has_weak_ref && list_empty(&node->work.entry)) {
			if (target_list == NULL) {
1093 1094
				pr_err("invalid inc weak node for %d\n",
					node->debug_id);
1095 1096
				return -EINVAL;
			}
1097
			binder_enqueue_work_ilocked(&node->work, target_list);
1098 1099 1100 1101 1102
		}
	}
	return 0;
}

1103 1104 1105 1106 1107
static int binder_inc_node(struct binder_node *node, int strong, int internal,
			   struct list_head *target_list)
{
	int ret;

1108 1109 1110
	binder_node_inner_lock(node);
	ret = binder_inc_node_nilocked(node, strong, internal, target_list);
	binder_node_inner_unlock(node);
1111 1112 1113 1114

	return ret;
}

1115 1116
static bool binder_dec_node_nilocked(struct binder_node *node,
				     int strong, int internal)
1117
{
1118 1119
	struct binder_proc *proc = node->proc;

1120
	BUG_ON(!spin_is_locked(&node->lock));
1121 1122
	if (proc)
		BUG_ON(!spin_is_locked(&proc->inner_lock));
1123 1124 1125 1126 1127 1128
	if (strong) {
		if (internal)
			node->internal_strong_refs--;
		else
			node->local_strong_refs--;
		if (node->local_strong_refs || node->internal_strong_refs)
1129
			return false;
1130 1131 1132
	} else {
		if (!internal)
			node->local_weak_refs--;
1133 1134
		if (node->local_weak_refs || node->tmp_refs ||
				!hlist_empty(&node->refs))
1135
			return false;
1136
	}
1137 1138

	if (proc && (node->has_strong_ref || node->has_weak_ref)) {
1139
		if (list_empty(&node->work.entry)) {
1140
			binder_enqueue_work_ilocked(&node->work, &proc->todo);
1141 1142 1143 1144
			wake_up_interruptible(&node->proc->wait);
		}
	} else {
		if (hlist_empty(&node->refs) && !node->local_strong_refs &&
1145
		    !node->local_weak_refs && !node->tmp_refs) {
1146
			if (proc) {
1147 1148
				binder_dequeue_work_ilocked(&node->work);
				rb_erase(&node->rb_node, &proc->nodes);
1149
				binder_debug(BINDER_DEBUG_INTERNAL_REFS,
1150
					     "refless node %d deleted\n",
1151 1152
					     node->debug_id);
			} else {
1153
				BUG_ON(!list_empty(&node->work.entry));
1154
				spin_lock(&binder_dead_nodes_lock);
1155 1156 1157 1158 1159 1160 1161 1162
				/*
				 * tmp_refs could have changed so
				 * check it again
				 */
				if (node->tmp_refs) {
					spin_unlock(&binder_dead_nodes_lock);
					return false;
				}
1163
				hlist_del(&node->dead_node);
1164
				spin_unlock(&binder_dead_nodes_lock);
1165
				binder_debug(BINDER_DEBUG_INTERNAL_REFS,
1166
					     "dead node %d deleted\n",
1167 1168
					     node->debug_id);
			}
1169
			return true;
1170 1171
		}
	}
1172 1173
	return false;
}
1174

1175 1176 1177 1178
static void binder_dec_node(struct binder_node *node, int strong, int internal)
{
	bool free_node;

1179 1180 1181
	binder_node_inner_lock(node);
	free_node = binder_dec_node_nilocked(node, strong, internal);
	binder_node_inner_unlock(node);
1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193
	if (free_node)
		binder_free_node(node);
}

static void binder_inc_node_tmpref_ilocked(struct binder_node *node)
{
	/*
	 * No call to binder_inc_node() is needed since we
	 * don't need to inform userspace of any changes to
	 * tmp_refs
	 */
	node->tmp_refs++;
1194 1195
}

1196 1197 1198 1199 1200
/**
 * binder_inc_node_tmpref() - take a temporary reference on node
 * @node:	node to reference
 *
 * Take reference on node to prevent the node from being freed
1201 1202 1203 1204 1205 1206 1207
 * while referenced only by a local variable. The inner lock is
 * needed to serialize with the node work on the queue (which
 * isn't needed after the node is dead). If the node is dead
 * (node->proc is NULL), use binder_dead_nodes_lock to protect
 * node->tmp_refs against dead-node-only cases where the node
 * lock cannot be acquired (eg traversing the dead node list to
 * print nodes)
1208 1209 1210
 */
static void binder_inc_node_tmpref(struct binder_node *node)
{
1211
	binder_node_lock(node);
1212 1213 1214 1215 1216 1217 1218 1219 1220
	if (node->proc)
		binder_inner_proc_lock(node->proc);
	else
		spin_lock(&binder_dead_nodes_lock);
	binder_inc_node_tmpref_ilocked(node);
	if (node->proc)
		binder_inner_proc_unlock(node->proc);
	else
		spin_unlock(&binder_dead_nodes_lock);
1221
	binder_node_unlock(node);
1222 1223 1224 1225 1226 1227 1228 1229 1230 1231
}

/**
 * binder_dec_node_tmpref() - remove a temporary reference on node
 * @node:	node to reference
 *
 * Release temporary reference on node taken via binder_inc_node_tmpref()
 */
static void binder_dec_node_tmpref(struct binder_node *node)
{
1232 1233
	bool free_node;

1234 1235
	binder_node_inner_lock(node);
	if (!node->proc)
1236
		spin_lock(&binder_dead_nodes_lock);
1237 1238
	node->tmp_refs--;
	BUG_ON(node->tmp_refs < 0);
1239 1240
	if (!node->proc)
		spin_unlock(&binder_dead_nodes_lock);
1241 1242 1243 1244 1245 1246
	/*
	 * Call binder_dec_node() to check if all refcounts are 0
	 * and cleanup is needed. Calling with strong=0 and internal=1
	 * causes no actual reference to be released in binder_dec_node().
	 * If that changes, a change is needed here too.
	 */
1247 1248
	free_node = binder_dec_node_nilocked(node, 0, 1);
	binder_node_inner_unlock(node);
1249 1250
	if (free_node)
		binder_free_node(node);
1251 1252 1253 1254 1255 1256
}

static void binder_put_node(struct binder_node *node)
{
	binder_dec_node_tmpref(node);
}
1257

1258 1259
static struct binder_ref *binder_get_ref_olocked(struct binder_proc *proc,
						 u32 desc, bool need_strong_ref)
1260 1261 1262 1263 1264 1265 1266
{
	struct rb_node *n = proc->refs_by_desc.rb_node;
	struct binder_ref *ref;

	while (n) {
		ref = rb_entry(n, struct binder_ref, rb_node_desc);

1267
		if (desc < ref->data.desc) {
1268
			n = n->rb_left;
1269
		} else if (desc > ref->data.desc) {
1270
			n = n->rb_right;
1271
		} else if (need_strong_ref && !ref->data.strong) {
1272 1273 1274
			binder_user_error("tried to use weak ref as strong ref\n");
			return NULL;
		} else {
1275
			return ref;
1276
		}
1277 1278 1279 1280
	}
	return NULL;
}

1281
/**
1282
 * binder_get_ref_for_node_olocked() - get the ref associated with given node
1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298
 * @proc:	binder_proc that owns the ref
 * @node:	binder_node of target
 * @new_ref:	newly allocated binder_ref to be initialized or %NULL
 *
 * Look up the ref for the given node and return it if it exists
 *
 * If it doesn't exist and the caller provides a newly allocated
 * ref, initialize the fields of the newly allocated ref and insert
 * into the given proc rb_trees and node refs list.
 *
 * Return:	the ref for node. It is possible that another thread
 *		allocated/initialized the ref first in which case the
 *		returned ref would be different than the passed-in
 *		new_ref. new_ref must be kfree'd by the caller in
 *		this case.
 */
1299 1300 1301 1302
static struct binder_ref *binder_get_ref_for_node_olocked(
					struct binder_proc *proc,
					struct binder_node *node,
					struct binder_ref *new_ref)
1303
{
1304
	struct binder_context *context = proc->context;
1305 1306
	struct rb_node **p = &proc->refs_by_node.rb_node;
	struct rb_node *parent = NULL;
1307 1308
	struct binder_ref *ref;
	struct rb_node *n;
1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320

	while (*p) {
		parent = *p;
		ref = rb_entry(parent, struct binder_ref, rb_node_node);

		if (node < ref->node)
			p = &(*p)->rb_left;
		else if (node > ref->node)
			p = &(*p)->rb_right;
		else
			return ref;
	}
1321
	if (!new_ref)
1322
		return NULL;
1323

1324
	binder_stats_created(BINDER_STAT_REF);
1325
	new_ref->data.debug_id = atomic_inc_return(&binder_last_id);
1326 1327 1328 1329 1330
	new_ref->proc = proc;
	new_ref->node = node;
	rb_link_node(&new_ref->rb_node_node, parent, p);
	rb_insert_color(&new_ref->rb_node_node, &proc->refs_by_node);

1331
	new_ref->data.desc = (node == context->binder_context_mgr_node) ? 0 : 1;
1332 1333
	for (n = rb_first(&proc->refs_by_desc); n != NULL; n = rb_next(n)) {
		ref = rb_entry(n, struct binder_ref, rb_node_desc);
1334
		if (ref->data.desc > new_ref->data.desc)
1335
			break;
1336
		new_ref->data.desc = ref->data.desc + 1;
1337 1338 1339 1340 1341 1342 1343
	}

	p = &proc->refs_by_desc.rb_node;
	while (*p) {
		parent = *p;
		ref = rb_entry(parent, struct binder_ref, rb_node_desc);

1344
		if (new_ref->data.desc < ref->data.desc)
1345
			p = &(*p)->rb_left;
1346
		else if (new_ref->data.desc > ref->data.desc)
1347 1348 1349 1350 1351 1352
			p = &(*p)->rb_right;
		else
			BUG();
	}
	rb_link_node(&new_ref->rb_node_desc, parent, p);
	rb_insert_color(&new_ref->rb_node_desc, &proc->refs_by_desc);
1353 1354

	binder_node_lock(node);
1355
	hlist_add_head(&new_ref->node_entry, &node->refs);
1356

1357 1358
	binder_debug(BINDER_DEBUG_INTERNAL_REFS,
		     "%d new ref %d desc %d for node %d\n",
1359
		      proc->pid, new_ref->data.debug_id, new_ref->data.desc,
1360
		      node->debug_id);
1361
	binder_node_unlock(node);
1362 1363 1364
	return new_ref;
}

1365
static void binder_cleanup_ref_olocked(struct binder_ref *ref)
1366
{
1367 1368
	bool delete_node = false;

1369
	binder_debug(BINDER_DEBUG_INTERNAL_REFS,
1370
		     "%d delete ref %d desc %d for node %d\n",
1371
		      ref->proc->pid, ref->data.debug_id, ref->data.desc,
1372
		      ref->node->debug_id);
1373 1374 1375

	rb_erase(&ref->rb_node_desc, &ref->proc->refs_by_desc);
	rb_erase(&ref->rb_node_node, &ref->proc->refs_by_node);
1376

1377
	binder_node_inner_lock(ref->node);
1378
	if (ref->data.strong)
1379
		binder_dec_node_nilocked(ref->node, 1, 1);
1380

1381
	hlist_del(&ref->node_entry);
1382 1383
	delete_node = binder_dec_node_nilocked(ref->node, 0, 1);
	binder_node_inner_unlock(ref->node);
1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394
	/*
	 * Clear ref->node unless we want the caller to free the node
	 */
	if (!delete_node) {
		/*
		 * The caller uses ref->node to determine
		 * whether the node needs to be freed. Clear
		 * it since the node is still alive.
		 */
		ref->node = NULL;
	}
1395

1396 1397
	if (ref->death) {
		binder_debug(BINDER_DEBUG_DEAD_BINDER,
1398
			     "%d delete ref %d desc %d has death notification\n",
1399 1400
			      ref->proc->pid, ref->data.debug_id,
			      ref->data.desc);
1401
		binder_dequeue_work(ref->proc, &ref->death->work);
1402 1403 1404 1405 1406
		binder_stats_deleted(BINDER_STAT_DEATH);
	}
	binder_stats_deleted(BINDER_STAT_REF);
}

1407
/**
1408
 * binder_inc_ref_olocked() - increment the ref for given handle
1409 1410 1411 1412
 * @ref:         ref to be incremented
 * @strong:      if true, strong increment, else weak
 * @target_list: list to queue node work on
 *
1413
 * Increment the ref. @ref->proc->outer_lock must be held on entry
1414 1415 1416
 *
 * Return: 0, if successful, else errno
 */
1417 1418
static int binder_inc_ref_olocked(struct binder_ref *ref, int strong,
				  struct list_head *target_list)
1419 1420
{
	int ret;
1421

1422
	if (strong) {
1423
		if (ref->data.strong == 0) {
1424 1425 1426 1427
			ret = binder_inc_node(ref->node, 1, 1, target_list);
			if (ret)
				return ret;
		}
1428
		ref->data.strong++;
1429
	} else {
1430
		if (ref->data.weak == 0) {
1431 1432 1433 1434
			ret = binder_inc_node(ref->node, 0, 1, target_list);
			if (ret)
				return ret;
		}
1435
		ref->data.weak++;
1436 1437 1438 1439
	}
	return 0;
}

1440 1441 1442 1443 1444 1445 1446 1447 1448
/**
 * binder_dec_ref() - dec the ref for given handle
 * @ref:	ref to be decremented
 * @strong:	if true, strong decrement, else weak
 *
 * Decrement the ref.
 *
 * Return: true if ref is cleaned up and ready to be freed
 */
1449
static bool binder_dec_ref_olocked(struct binder_ref *ref, int strong)
1450 1451
{
	if (strong) {
1452
		if (ref->data.strong == 0) {
1453
			binder_user_error("%d invalid dec strong, ref %d desc %d s %d w %d\n",
1454 1455 1456 1457
					  ref->proc->pid, ref->data.debug_id,
					  ref->data.desc, ref->data.strong,
					  ref->data.weak);
			return false;
1458
		}
1459
		ref->data.strong--;
1460 1461
		if (ref->data.strong == 0)
			binder_dec_node(ref->node, strong, 1);
1462
	} else {
1463
		if (ref->data.weak == 0) {
1464
			binder_user_error("%d invalid dec weak, ref %d desc %d s %d w %d\n",
1465 1466 1467 1468
					  ref->proc->pid, ref->data.debug_id,
					  ref->data.desc, ref->data.strong,
					  ref->data.weak);
			return false;
1469
		}
1470
		ref->data.weak--;
1471
	}
1472
	if (ref->data.strong == 0 && ref->data.weak == 0) {
1473
		binder_cleanup_ref_olocked(ref);
1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497
		return true;
	}
	return false;
}

/**
 * binder_get_node_from_ref() - get the node from the given proc/desc
 * @proc:	proc containing the ref
 * @desc:	the handle associated with the ref
 * @need_strong_ref: if true, only return node if ref is strong
 * @rdata:	the id/refcount data for the ref
 *
 * Given a proc and ref handle, return the associated binder_node
 *
 * Return: a binder_node or NULL if not found or not strong when strong required
 */
static struct binder_node *binder_get_node_from_ref(
		struct binder_proc *proc,
		u32 desc, bool need_strong_ref,
		struct binder_ref_data *rdata)
{
	struct binder_node *node;
	struct binder_ref *ref;

1498 1499
	binder_proc_lock(proc);
	ref = binder_get_ref_olocked(proc, desc, need_strong_ref);
1500 1501 1502
	if (!ref)
		goto err_no_ref;
	node = ref->node;
1503 1504 1505 1506 1507
	/*
	 * Take an implicit reference on the node to ensure
	 * it stays alive until the call to binder_put_node()
	 */
	binder_inc_node_tmpref(node);
1508 1509
	if (rdata)
		*rdata = ref->data;
1510
	binder_proc_unlock(proc);
1511 1512 1513 1514

	return node;

err_no_ref:
1515
	binder_proc_unlock(proc);
1516 1517 1518 1519 1520 1521 1522
	return NULL;
}

/**
 * binder_free_ref() - free the binder_ref
 * @ref:	ref to free
 *
1523 1524
 * Free the binder_ref. Free the binder_node indicated by ref->node
 * (if non-NULL) and the binder_ref_death indicated by ref->death.
1525 1526 1527
 */
static void binder_free_ref(struct binder_ref *ref)
{
1528 1529
	if (ref->node)
		binder_free_node(ref->node);
1530 1531 1532 1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554
	kfree(ref->death);
	kfree(ref);
}

/**
 * binder_update_ref_for_handle() - inc/dec the ref for given handle
 * @proc:	proc containing the ref
 * @desc:	the handle associated with the ref
 * @increment:	true=inc reference, false=dec reference
 * @strong:	true=strong reference, false=weak reference
 * @rdata:	the id/refcount data for the ref
 *
 * Given a proc and ref handle, increment or decrement the ref
 * according to "increment" arg.
 *
 * Return: 0 if successful, else errno
 */
static int binder_update_ref_for_handle(struct binder_proc *proc,
		uint32_t desc, bool increment, bool strong,
		struct binder_ref_data *rdata)
{
	int ret = 0;
	struct binder_ref *ref;
	bool delete_ref = false;

1555 1556
	binder_proc_lock(proc);
	ref = binder_get_ref_olocked(proc, desc, strong);
1557 1558 1559 1560 1561
	if (!ref) {
		ret = -EINVAL;
		goto err_no_ref;
	}
	if (increment)
1562
		ret = binder_inc_ref_olocked(ref, strong, NULL);
1563
	else
1564
		delete_ref = binder_dec_ref_olocked(ref, strong);
1565 1566 1567

	if (rdata)
		*rdata = ref->data;
1568
	binder_proc_unlock(proc);
1569 1570 1571 1572 1573 1574

	if (delete_ref)
		binder_free_ref(ref);
	return ret;

err_no_ref:
1575
	binder_proc_unlock(proc);
1576 1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619
	return ret;
}

/**
 * binder_dec_ref_for_handle() - dec the ref for given handle
 * @proc:	proc containing the ref
 * @desc:	the handle associated with the ref
 * @strong:	true=strong reference, false=weak reference
 * @rdata:	the id/refcount data for the ref
 *
 * Just calls binder_update_ref_for_handle() to decrement the ref.
 *
 * Return: 0 if successful, else errno
 */
static int binder_dec_ref_for_handle(struct binder_proc *proc,
		uint32_t desc, bool strong, struct binder_ref_data *rdata)
{
	return binder_update_ref_for_handle(proc, desc, false, strong, rdata);
}


/**
 * binder_inc_ref_for_node() - increment the ref for given proc/node
 * @proc:	 proc containing the ref
 * @node:	 target node
 * @strong:	 true=strong reference, false=weak reference
 * @target_list: worklist to use if node is incremented
 * @rdata:	 the id/refcount data for the ref
 *
 * Given a proc and node, increment the ref. Create the ref if it
 * doesn't already exist
 *
 * Return: 0 if successful, else errno
 */
static int binder_inc_ref_for_node(struct binder_proc *proc,
			struct binder_node *node,
			bool strong,
			struct list_head *target_list,
			struct binder_ref_data *rdata)
{
	struct binder_ref *ref;
	struct binder_ref *new_ref = NULL;
	int ret = 0;

1620 1621
	binder_proc_lock(proc);
	ref = binder_get_ref_for_node_olocked(proc, node, NULL);
1622
	if (!ref) {
1623
		binder_proc_unlock(proc);
1624 1625 1626
		new_ref = kzalloc(sizeof(*ref), GFP_KERNEL);
		if (!new_ref)
			return -ENOMEM;
1627 1628
		binder_proc_lock(proc);
		ref = binder_get_ref_for_node_olocked(proc, node, new_ref);
1629
	}
1630
	ret = binder_inc_ref_olocked(ref, strong, target_list);
1631
	*rdata = ref->data;
1632
	binder_proc_unlock(proc);
1633 1634 1635 1636 1637 1638 1639
	if (new_ref && ref != new_ref)
		/*
		 * Another thread created the ref first so
		 * free the one we allocated
		 */
		kfree(new_ref);
	return ret;
1640 1641
}

1642 1643
static void binder_pop_transaction_ilocked(struct binder_thread *target_thread,
					   struct binder_transaction *t)
1644
{
1645
	BUG_ON(!target_thread);
1646
	BUG_ON(!spin_is_locked(&target_thread->proc->inner_lock));
1647 1648 1649 1650 1651 1652 1653
	BUG_ON(target_thread->transaction_stack != t);
	BUG_ON(target_thread->transaction_stack->from != target_thread);
	target_thread->transaction_stack =
		target_thread->transaction_stack->from_parent;
	t->from = NULL;
}

1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671
/**
 * binder_thread_dec_tmpref() - decrement thread->tmp_ref
 * @thread:	thread to decrement
 *
 * A thread needs to be kept alive while being used to create or
 * handle a transaction. binder_get_txn_from() is used to safely
 * extract t->from from a binder_transaction and keep the thread
 * indicated by t->from from being freed. When done with that
 * binder_thread, this function is called to decrement the
 * tmp_ref and free if appropriate (thread has been released
 * and no transaction being processed by the driver)
 */
static void binder_thread_dec_tmpref(struct binder_thread *thread)
{
	/*
	 * atomic is used to protect the counter value while
	 * it cannot reach zero or thread->is_dead is false
	 */
1672
	binder_inner_proc_lock(thread->proc);
1673 1674
	atomic_dec(&thread->tmp_ref);
	if (thread->is_dead && !atomic_read(&thread->tmp_ref)) {
1675
		binder_inner_proc_unlock(thread->proc);
1676 1677 1678
		binder_free_thread(thread);
		return;
	}
1679
	binder_inner_proc_unlock(thread->proc);
1680 1681 1682 1683 1684 1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695
}

/**
 * binder_proc_dec_tmpref() - decrement proc->tmp_ref
 * @proc:	proc to decrement
 *
 * A binder_proc needs to be kept alive while being used to create or
 * handle a transaction. proc->tmp_ref is incremented when
 * creating a new transaction or the binder_proc is currently in-use
 * by threads that are being released. When done with the binder_proc,
 * this function is called to decrement the counter and free the
 * proc if appropriate (proc has been released, all threads have
 * been released and not currenly in-use to process a transaction).
 */
static void binder_proc_dec_tmpref(struct binder_proc *proc)
{
1696
	binder_inner_proc_lock(proc);
1697 1698 1699
	proc->tmp_ref--;
	if (proc->is_dead && RB_EMPTY_ROOT(&proc->threads) &&
			!proc->tmp_ref) {
1700
		binder_inner_proc_unlock(proc);
1701 1702 1703
		binder_free_proc(proc);
		return;
	}
1704
	binder_inner_proc_unlock(proc);
1705 1706 1707 1708 1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726 1727 1728 1729
}

/**
 * binder_get_txn_from() - safely extract the "from" thread in transaction
 * @t:	binder transaction for t->from
 *
 * Atomically return the "from" thread and increment the tmp_ref
 * count for the thread to ensure it stays alive until
 * binder_thread_dec_tmpref() is called.
 *
 * Return: the value of t->from
 */
static struct binder_thread *binder_get_txn_from(
		struct binder_transaction *t)
{
	struct binder_thread *from;

	spin_lock(&t->lock);
	from = t->from;
	if (from)
		atomic_inc(&from->tmp_ref);
	spin_unlock(&t->lock);
	return from;
}

1730 1731 1732 1733 1734 1735 1736 1737 1738 1739 1740 1741 1742 1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753 1754 1755 1756 1757 1758
/**
 * binder_get_txn_from_and_acq_inner() - get t->from and acquire inner lock
 * @t:	binder transaction for t->from
 *
 * Same as binder_get_txn_from() except it also acquires the proc->inner_lock
 * to guarantee that the thread cannot be released while operating on it.
 * The caller must call binder_inner_proc_unlock() to release the inner lock
 * as well as call binder_dec_thread_txn() to release the reference.
 *
 * Return: the value of t->from
 */
static struct binder_thread *binder_get_txn_from_and_acq_inner(
		struct binder_transaction *t)
{
	struct binder_thread *from;

	from = binder_get_txn_from(t);
	if (!from)
		return NULL;
	binder_inner_proc_lock(from->proc);
	if (t->from) {
		BUG_ON(from != t->from);
		return from;
	}
	binder_inner_proc_unlock(from->proc);
	binder_thread_dec_tmpref(from);
	return NULL;
}

1759 1760
static void binder_free_transaction(struct binder_transaction *t)
{
1761 1762 1763 1764 1765 1766 1767 1768 1769 1770
	if (t->buffer)
		t->buffer->transaction = NULL;
	kfree(t);
	binder_stats_deleted(BINDER_STAT_TRANSACTION);
}

static void binder_send_failed_reply(struct binder_transaction *t,
				     uint32_t error_code)
{
	struct binder_thread *target_thread;
1771
	struct binder_transaction *next;
1772

1773 1774
	BUG_ON(t->flags & TF_ONE_WAY);
	while (1) {
1775
		target_thread = binder_get_txn_from_and_acq_inner(t);
1776
		if (target_thread) {
1777 1778 1779 1780 1781 1782
			binder_debug(BINDER_DEBUG_FAILED_TRANSACTION,
				     "send failed reply for transaction %d to %d:%d\n",
				      t->debug_id,
				      target_thread->proc->pid,
				      target_thread->pid);

1783
			binder_pop_transaction_ilocked(target_thread, t);
1784 1785
			if (target_thread->reply_error.cmd == BR_OK) {
				target_thread->reply_error.cmd = error_code;
1786
				binder_enqueue_work_ilocked(
1787
					&target_thread->reply_error.work,
1788
					&target_thread->todo);
1789 1790
				wake_up_interruptible(&target_thread->wait);
			} else {
1791 1792
				WARN(1, "Unexpected reply error: %u\n",
						target_thread->reply_error.cmd);
1793
			}
1794
			binder_inner_proc_unlock(target_thread->proc);
1795
			binder_thread_dec_tmpref(target_thread);
1796
			binder_free_transaction(t);
1797
			return;
1798 1799 1800 1801 1802 1803 1804
		}
		next = t->from_parent;

		binder_debug(BINDER_DEBUG_FAILED_TRANSACTION,
			     "send failed reply for transaction %d, target dead\n",
			     t->debug_id);

1805
		binder_free_transaction(t);
1806
		if (next == NULL) {
1807
			binder_debug(BINDER_DEBUG_DEAD_BINDER,
1808 1809
				     "reply failed, no target thread at root\n");
			return;
1810
		}
1811 1812 1813 1814
		t = next;
		binder_debug(BINDER_DEBUG_DEAD_BINDER,
			     "reply failed, no target thread -- retry %d\n",
			      t->debug_id);
1815 1816 1817
	}
}

1818 1819 1820 1821 1822 1823 1824 1825 1826 1827 1828 1829 1830 1831 1832 1833 1834 1835 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845 1846 1847 1848
/**
 * binder_validate_object() - checks for a valid metadata object in a buffer.
 * @buffer:	binder_buffer that we're parsing.
 * @offset:	offset in the buffer at which to validate an object.
 *
 * Return:	If there's a valid metadata object at @offset in @buffer, the
 *		size of that object. Otherwise, it returns zero.
 */
static size_t binder_validate_object(struct binder_buffer *buffer, u64 offset)
{
	/* Check if we can read a header first */
	struct binder_object_header *hdr;
	size_t object_size = 0;

	if (offset > buffer->data_size - sizeof(*hdr) ||
	    buffer->data_size < sizeof(*hdr) ||
	    !IS_ALIGNED(offset, sizeof(u32)))
		return 0;

	/* Ok, now see if we can read a complete object. */
	hdr = (struct binder_object_header *)(buffer->data + offset);
	switch (hdr->type) {
	case BINDER_TYPE_BINDER:
	case BINDER_TYPE_WEAK_BINDER:
	case BINDER_TYPE_HANDLE:
	case BINDER_TYPE_WEAK_HANDLE:
		object_size = sizeof(struct flat_binder_object);
		break;
	case BINDER_TYPE_FD:
		object_size = sizeof(struct binder_fd_object);
		break;
1849 1850 1851
	case BINDER_TYPE_PTR:
		object_size = sizeof(struct binder_buffer_object);
		break;
1852 1853 1854
	case BINDER_TYPE_FDA:
		object_size = sizeof(struct binder_fd_array_object);
		break;
1855 1856 1857 1858 1859 1860 1861 1862 1863 1864
	default:
		return 0;
	}
	if (offset <= buffer->data_size - object_size &&
	    buffer->data_size >= object_size)
		return object_size;
	else
		return 0;
}

1865 1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880 1881 1882 1883 1884 1885 1886 1887 1888 1889 1890 1891 1892 1893 1894 1895 1896 1897 1898 1899 1900 1901 1902 1903 1904 1905 1906 1907 1908 1909 1910 1911 1912 1913 1914 1915 1916 1917 1918 1919 1920 1921 1922 1923 1924 1925 1926 1927 1928 1929 1930 1931 1932 1933 1934 1935 1936 1937 1938 1939 1940 1941 1942 1943 1944 1945 1946 1947 1948 1949 1950 1951 1952 1953 1954 1955 1956 1957 1958 1959 1960 1961 1962 1963 1964
/**
 * binder_validate_ptr() - validates binder_buffer_object in a binder_buffer.
 * @b:		binder_buffer containing the object
 * @index:	index in offset array at which the binder_buffer_object is
 *		located
 * @start:	points to the start of the offset array
 * @num_valid:	the number of valid offsets in the offset array
 *
 * Return:	If @index is within the valid range of the offset array
 *		described by @start and @num_valid, and if there's a valid
 *		binder_buffer_object at the offset found in index @index
 *		of the offset array, that object is returned. Otherwise,
 *		%NULL is returned.
 *		Note that the offset found in index @index itself is not
 *		verified; this function assumes that @num_valid elements
 *		from @start were previously verified to have valid offsets.
 */
static struct binder_buffer_object *binder_validate_ptr(struct binder_buffer *b,
							binder_size_t index,
							binder_size_t *start,
							binder_size_t num_valid)
{
	struct binder_buffer_object *buffer_obj;
	binder_size_t *offp;

	if (index >= num_valid)
		return NULL;

	offp = start + index;
	buffer_obj = (struct binder_buffer_object *)(b->data + *offp);
	if (buffer_obj->hdr.type != BINDER_TYPE_PTR)
		return NULL;

	return buffer_obj;
}

/**
 * binder_validate_fixup() - validates pointer/fd fixups happen in order.
 * @b:			transaction buffer
 * @objects_start	start of objects buffer
 * @buffer:		binder_buffer_object in which to fix up
 * @offset:		start offset in @buffer to fix up
 * @last_obj:		last binder_buffer_object that we fixed up in
 * @last_min_offset:	minimum fixup offset in @last_obj
 *
 * Return:		%true if a fixup in buffer @buffer at offset @offset is
 *			allowed.
 *
 * For safety reasons, we only allow fixups inside a buffer to happen
 * at increasing offsets; additionally, we only allow fixup on the last
 * buffer object that was verified, or one of its parents.
 *
 * Example of what is allowed:
 *
 * A
 *   B (parent = A, offset = 0)
 *   C (parent = A, offset = 16)
 *     D (parent = C, offset = 0)
 *   E (parent = A, offset = 32) // min_offset is 16 (C.parent_offset)
 *
 * Examples of what is not allowed:
 *
 * Decreasing offsets within the same parent:
 * A
 *   C (parent = A, offset = 16)
 *   B (parent = A, offset = 0) // decreasing offset within A
 *
 * Referring to a parent that wasn't the last object or any of its parents:
 * A
 *   B (parent = A, offset = 0)
 *   C (parent = A, offset = 0)
 *   C (parent = A, offset = 16)
 *     D (parent = B, offset = 0) // B is not A or any of A's parents
 */
static bool binder_validate_fixup(struct binder_buffer *b,
				  binder_size_t *objects_start,
				  struct binder_buffer_object *buffer,
				  binder_size_t fixup_offset,
				  struct binder_buffer_object *last_obj,
				  binder_size_t last_min_offset)
{
	if (!last_obj) {
		/* Nothing to fix up in */
		return false;
	}

	while (last_obj != buffer) {
		/*
		 * Safe to retrieve the parent of last_obj, since it
		 * was already previously verified by the driver.
		 */
		if ((last_obj->flags & BINDER_BUFFER_FLAG_HAS_PARENT) == 0)
			return false;
		last_min_offset = last_obj->parent_offset + sizeof(uintptr_t);
		last_obj = (struct binder_buffer_object *)
			(b->data + *(objects_start + last_obj->parent));
	}
	return (fixup_offset >= last_min_offset);
}

1965 1966
static void binder_transaction_buffer_release(struct binder_proc *proc,
					      struct binder_buffer *buffer,
1967
					      binder_size_t *failed_at)
1968
{
1969
	binder_size_t *offp, *off_start, *off_end;
1970 1971 1972
	int debug_id = buffer->debug_id;

	binder_debug(BINDER_DEBUG_TRANSACTION,
1973
		     "%d buffer release %d, size %zd-%zd, failed at %p\n",
1974 1975 1976 1977 1978 1979
		     proc->pid, buffer->debug_id,
		     buffer->data_size, buffer->offsets_size, failed_at);

	if (buffer->target_node)
		binder_dec_node(buffer->target_node, 1, 0);

1980 1981
	off_start = (binder_size_t *)(buffer->data +
				      ALIGN(buffer->data_size, sizeof(void *)));
1982 1983 1984
	if (failed_at)
		off_end = failed_at;
	else
1985 1986
		off_end = (void *)off_start + buffer->offsets_size;
	for (offp = off_start; offp < off_end; offp++) {
1987 1988
		struct binder_object_header *hdr;
		size_t object_size = binder_validate_object(buffer, *offp);
1989

1990 1991
		if (object_size == 0) {
			pr_err("transaction release %d bad object at offset %lld, size %zd\n",
1992
			       debug_id, (u64)*offp, buffer->data_size);
1993 1994
			continue;
		}
1995 1996
		hdr = (struct binder_object_header *)(buffer->data + *offp);
		switch (hdr->type) {
1997 1998
		case BINDER_TYPE_BINDER:
		case BINDER_TYPE_WEAK_BINDER: {
1999 2000
			struct flat_binder_object *fp;
			struct binder_node *node;
2001

2002 2003
			fp = to_flat_binder_object(hdr);
			node = binder_get_node(proc, fp->binder);
2004
			if (node == NULL) {
2005 2006
				pr_err("transaction release %d bad node %016llx\n",
				       debug_id, (u64)fp->binder);
2007 2008 2009
				break;
			}
			binder_debug(BINDER_DEBUG_TRANSACTION,
2010 2011
				     "        node %d u%016llx\n",
				     node->debug_id, (u64)node->ptr);
2012 2013
			binder_dec_node(node, hdr->type == BINDER_TYPE_BINDER,
					0);
2014
			binder_put_node(node);
2015 2016 2017
		} break;
		case BINDER_TYPE_HANDLE:
		case BINDER_TYPE_WEAK_HANDLE: {
2018
			struct flat_binder_object *fp;
2019 2020
			struct binder_ref_data rdata;
			int ret;
2021

2022
			fp = to_flat_binder_object(hdr);
2023 2024 2025 2026 2027 2028
			ret = binder_dec_ref_for_handle(proc, fp->handle,
				hdr->type == BINDER_TYPE_HANDLE, &rdata);

			if (ret) {
				pr_err("transaction release %d bad handle %d, ret = %d\n",
				 debug_id, fp->handle, ret);
2029 2030 2031
				break;
			}
			binder_debug(BINDER_DEBUG_TRANSACTION,
2032 2033
				     "        ref %d desc %d\n",
				     rdata.debug_id, rdata.desc);
2034 2035
		} break;

2036 2037 2038
		case BINDER_TYPE_FD: {
			struct binder_fd_object *fp = to_binder_fd_object(hdr);

2039
			binder_debug(BINDER_DEBUG_TRANSACTION,
2040
				     "        fd %d\n", fp->fd);
2041
			if (failed_at)
2042 2043
				task_close_fd(proc, fp->fd);
		} break;
2044 2045 2046 2047 2048 2049
		case BINDER_TYPE_PTR:
			/*
			 * Nothing to do here, this will get cleaned up when the
			 * transaction buffer gets freed
			 */
			break;
2050 2051 2052 2053 2054 2055 2056 2057 2058 2059 2060 2061 2062 2063 2064 2065 2066 2067 2068 2069 2070 2071
		case BINDER_TYPE_FDA: {
			struct binder_fd_array_object *fda;
			struct binder_buffer_object *parent;
			uintptr_t parent_buffer;
			u32 *fd_array;
			size_t fd_index;
			binder_size_t fd_buf_size;

			fda = to_binder_fd_array_object(hdr);
			parent = binder_validate_ptr(buffer, fda->parent,
						     off_start,
						     offp - off_start);
			if (!parent) {
				pr_err("transaction release %d bad parent offset",
				       debug_id);
				continue;
			}
			/*
			 * Since the parent was already fixed up, convert it
			 * back to kernel address space to access it
			 */
			parent_buffer = parent->buffer -
2072 2073
				binder_alloc_get_user_buffer_offset(
						&proc->alloc);
2074 2075 2076 2077 2078 2079 2080 2081 2082 2083 2084 2085 2086 2087 2088 2089 2090 2091

			fd_buf_size = sizeof(u32) * fda->num_fds;
			if (fda->num_fds >= SIZE_MAX / sizeof(u32)) {
				pr_err("transaction release %d invalid number of fds (%lld)\n",
				       debug_id, (u64)fda->num_fds);
				continue;
			}
			if (fd_buf_size > parent->length ||
			    fda->parent_offset > parent->length - fd_buf_size) {
				/* No space for all file descriptors here. */
				pr_err("transaction release %d not enough space for %lld fds in buffer\n",
				       debug_id, (u64)fda->num_fds);
				continue;
			}
			fd_array = (u32 *)(parent_buffer + fda->parent_offset);
			for (fd_index = 0; fd_index < fda->num_fds; fd_index++)
				task_close_fd(proc, fd_array[fd_index]);
		} break;
2092
		default:
2093
			pr_err("transaction release %d bad object type %x\n",
2094
				debug_id, hdr->type);
2095 2096 2097 2098 2099
			break;
		}
	}
}

2100 2101 2102 2103 2104 2105 2106
static int binder_translate_binder(struct flat_binder_object *fp,
				   struct binder_transaction *t,
				   struct binder_thread *thread)
{
	struct binder_node *node;
	struct binder_proc *proc = thread->proc;
	struct binder_proc *target_proc = t->to_proc;
2107
	struct binder_ref_data rdata;
2108
	int ret = 0;
2109 2110 2111

	node = binder_get_node(proc, fp->binder);
	if (!node) {
2112
		node = binder_new_node(proc, fp);
2113 2114 2115 2116 2117 2118 2119 2120
		if (!node)
			return -ENOMEM;
	}
	if (fp->cookie != node->cookie) {
		binder_user_error("%d:%d sending u%016llx node %d, cookie mismatch %016llx != %016llx\n",
				  proc->pid, thread->pid, (u64)fp->binder,
				  node->debug_id, (u64)fp->cookie,
				  (u64)node->cookie);
2121 2122 2123 2124 2125 2126
		ret = -EINVAL;
		goto done;
	}
	if (security_binder_transfer_binder(proc->tsk, target_proc->tsk)) {
		ret = -EPERM;
		goto done;
2127 2128
	}

2129 2130 2131 2132
	ret = binder_inc_ref_for_node(target_proc, node,
			fp->hdr.type == BINDER_TYPE_BINDER,
			&thread->todo, &rdata);
	if (ret)
2133
		goto done;
2134 2135 2136 2137 2138 2139

	if (fp->hdr.type == BINDER_TYPE_BINDER)
		fp->hdr.type = BINDER_TYPE_HANDLE;
	else
		fp->hdr.type = BINDER_TYPE_WEAK_HANDLE;
	fp->binder = 0;
2140
	fp->handle = rdata.desc;
2141 2142
	fp->cookie = 0;

2143
	trace_binder_transaction_node_to_ref(t, node, &rdata);
2144 2145 2146
	binder_debug(BINDER_DEBUG_TRANSACTION,
		     "        node %d u%016llx -> ref %d desc %d\n",
		     node->debug_id, (u64)node->ptr,
2147
		     rdata.debug_id, rdata.desc);
2148 2149 2150
done:
	binder_put_node(node);
	return ret;
2151 2152 2153 2154 2155 2156 2157 2158
}

static int binder_translate_handle(struct flat_binder_object *fp,
				   struct binder_transaction *t,
				   struct binder_thread *thread)
{
	struct binder_proc *proc = thread->proc;
	struct binder_proc *target_proc = t->to_proc;
2159 2160
	struct binder_node *node;
	struct binder_ref_data src_rdata;
2161
	int ret = 0;
2162

2163 2164 2165
	node = binder_get_node_from_ref(proc, fp->handle,
			fp->hdr.type == BINDER_TYPE_HANDLE, &src_rdata);
	if (!node) {
2166 2167 2168 2169
		binder_user_error("%d:%d got transaction with invalid handle, %d\n",
				  proc->pid, thread->pid, fp->handle);
		return -EINVAL;
	}
2170 2171 2172 2173
	if (security_binder_transfer_binder(proc->tsk, target_proc->tsk)) {
		ret = -EPERM;
		goto done;
	}
2174

2175
	binder_node_lock(node);
2176
	if (node->proc == target_proc) {
2177 2178 2179 2180
		if (fp->hdr.type == BINDER_TYPE_HANDLE)
			fp->hdr.type = BINDER_TYPE_BINDER;
		else
			fp->hdr.type = BINDER_TYPE_WEAK_BINDER;
2181 2182
		fp->binder = node->ptr;
		fp->cookie = node->cookie;
2183 2184 2185 2186 2187 2188 2189
		if (node->proc)
			binder_inner_proc_lock(node->proc);
		binder_inc_node_nilocked(node,
					 fp->hdr.type == BINDER_TYPE_BINDER,
					 0, NULL);
		if (node->proc)
			binder_inner_proc_unlock(node->proc);
2190
		trace_binder_transaction_ref_to_node(t, node, &src_rdata);
2191 2192
		binder_debug(BINDER_DEBUG_TRANSACTION,
			     "        ref %d desc %d -> node %d u%016llx\n",
2193 2194
			     src_rdata.debug_id, src_rdata.desc, node->debug_id,
			     (u64)node->ptr);
2195
		binder_node_unlock(node);
2196
	} else {
2197 2198
		int ret;
		struct binder_ref_data dest_rdata;
2199

2200
		binder_node_unlock(node);
2201 2202 2203 2204
		ret = binder_inc_ref_for_node(target_proc, node,
				fp->hdr.type == BINDER_TYPE_HANDLE,
				NULL, &dest_rdata);
		if (ret)
2205
			goto done;
2206 2207

		fp->binder = 0;
2208
		fp->handle = dest_rdata.desc;
2209
		fp->cookie = 0;
2210 2211
		trace_binder_transaction_ref_to_ref(t, node, &src_rdata,
						    &dest_rdata);
2212 2213
		binder_debug(BINDER_DEBUG_TRANSACTION,
			     "        ref %d desc %d -> ref %d desc %d (node %d)\n",
2214 2215 2216
			     src_rdata.debug_id, src_rdata.desc,
			     dest_rdata.debug_id, dest_rdata.desc,
			     node->debug_id);
2217
	}
2218 2219 2220
done:
	binder_put_node(node);
	return ret;
2221 2222 2223 2224 2225 2226 2227 2228 2229 2230 2231 2232 2233 2234 2235 2236 2237 2238 2239 2240 2241 2242 2243 2244 2245 2246 2247 2248 2249 2250 2251 2252 2253 2254 2255 2256 2257 2258 2259 2260 2261 2262 2263 2264 2265 2266 2267 2268 2269 2270 2271 2272 2273 2274 2275 2276 2277 2278 2279 2280
}

static int binder_translate_fd(int fd,
			       struct binder_transaction *t,
			       struct binder_thread *thread,
			       struct binder_transaction *in_reply_to)
{
	struct binder_proc *proc = thread->proc;
	struct binder_proc *target_proc = t->to_proc;
	int target_fd;
	struct file *file;
	int ret;
	bool target_allows_fd;

	if (in_reply_to)
		target_allows_fd = !!(in_reply_to->flags & TF_ACCEPT_FDS);
	else
		target_allows_fd = t->buffer->target_node->accept_fds;
	if (!target_allows_fd) {
		binder_user_error("%d:%d got %s with fd, %d, but target does not allow fds\n",
				  proc->pid, thread->pid,
				  in_reply_to ? "reply" : "transaction",
				  fd);
		ret = -EPERM;
		goto err_fd_not_accepted;
	}

	file = fget(fd);
	if (!file) {
		binder_user_error("%d:%d got transaction with invalid fd, %d\n",
				  proc->pid, thread->pid, fd);
		ret = -EBADF;
		goto err_fget;
	}
	ret = security_binder_transfer_file(proc->tsk, target_proc->tsk, file);
	if (ret < 0) {
		ret = -EPERM;
		goto err_security;
	}

	target_fd = task_get_unused_fd_flags(target_proc, O_CLOEXEC);
	if (target_fd < 0) {
		ret = -ENOMEM;
		goto err_get_unused_fd;
	}
	task_fd_install(target_proc, target_fd, file);
	trace_binder_transaction_fd(t, fd, target_fd);
	binder_debug(BINDER_DEBUG_TRANSACTION, "        fd %d -> %d\n",
		     fd, target_fd);

	return target_fd;

err_get_unused_fd:
err_security:
	fput(file);
err_fget:
err_fd_not_accepted:
	return ret;
}

2281 2282 2283 2284 2285 2286 2287 2288 2289 2290 2291 2292 2293 2294 2295 2296 2297 2298 2299 2300 2301 2302 2303 2304 2305 2306 2307 2308 2309 2310
static int binder_translate_fd_array(struct binder_fd_array_object *fda,
				     struct binder_buffer_object *parent,
				     struct binder_transaction *t,
				     struct binder_thread *thread,
				     struct binder_transaction *in_reply_to)
{
	binder_size_t fdi, fd_buf_size, num_installed_fds;
	int target_fd;
	uintptr_t parent_buffer;
	u32 *fd_array;
	struct binder_proc *proc = thread->proc;
	struct binder_proc *target_proc = t->to_proc;

	fd_buf_size = sizeof(u32) * fda->num_fds;
	if (fda->num_fds >= SIZE_MAX / sizeof(u32)) {
		binder_user_error("%d:%d got transaction with invalid number of fds (%lld)\n",
				  proc->pid, thread->pid, (u64)fda->num_fds);
		return -EINVAL;
	}
	if (fd_buf_size > parent->length ||
	    fda->parent_offset > parent->length - fd_buf_size) {
		/* No space for all file descriptors here. */
		binder_user_error("%d:%d not enough space to store %lld fds in buffer\n",
				  proc->pid, thread->pid, (u64)fda->num_fds);
		return -EINVAL;
	}
	/*
	 * Since the parent was already fixed up, convert it
	 * back to the kernel address space to access it
	 */
2311 2312
	parent_buffer = parent->buffer -
		binder_alloc_get_user_buffer_offset(&target_proc->alloc);
2313 2314 2315 2316 2317 2318 2319 2320 2321 2322 2323 2324 2325 2326 2327 2328 2329 2330 2331 2332 2333 2334 2335 2336 2337 2338
	fd_array = (u32 *)(parent_buffer + fda->parent_offset);
	if (!IS_ALIGNED((unsigned long)fd_array, sizeof(u32))) {
		binder_user_error("%d:%d parent offset not aligned correctly.\n",
				  proc->pid, thread->pid);
		return -EINVAL;
	}
	for (fdi = 0; fdi < fda->num_fds; fdi++) {
		target_fd = binder_translate_fd(fd_array[fdi], t, thread,
						in_reply_to);
		if (target_fd < 0)
			goto err_translate_fd_failed;
		fd_array[fdi] = target_fd;
	}
	return 0;

err_translate_fd_failed:
	/*
	 * Failed to allocate fd or security error, free fds
	 * installed so far.
	 */
	num_installed_fds = fdi;
	for (fdi = 0; fdi < num_installed_fds; fdi++)
		task_close_fd(target_proc, fd_array[fdi]);
	return target_fd;
}

2339 2340 2341 2342 2343 2344 2345 2346 2347 2348 2349 2350 2351 2352 2353 2354 2355 2356 2357 2358 2359 2360 2361 2362 2363 2364 2365 2366 2367 2368 2369 2370 2371 2372 2373 2374 2375 2376 2377 2378 2379
static int binder_fixup_parent(struct binder_transaction *t,
			       struct binder_thread *thread,
			       struct binder_buffer_object *bp,
			       binder_size_t *off_start,
			       binder_size_t num_valid,
			       struct binder_buffer_object *last_fixup_obj,
			       binder_size_t last_fixup_min_off)
{
	struct binder_buffer_object *parent;
	u8 *parent_buffer;
	struct binder_buffer *b = t->buffer;
	struct binder_proc *proc = thread->proc;
	struct binder_proc *target_proc = t->to_proc;

	if (!(bp->flags & BINDER_BUFFER_FLAG_HAS_PARENT))
		return 0;

	parent = binder_validate_ptr(b, bp->parent, off_start, num_valid);
	if (!parent) {
		binder_user_error("%d:%d got transaction with invalid parent offset or type\n",
				  proc->pid, thread->pid);
		return -EINVAL;
	}

	if (!binder_validate_fixup(b, off_start,
				   parent, bp->parent_offset,
				   last_fixup_obj,
				   last_fixup_min_off)) {
		binder_user_error("%d:%d got transaction with out-of-order buffer fixup\n",
				  proc->pid, thread->pid);
		return -EINVAL;
	}

	if (parent->length < sizeof(binder_uintptr_t) ||
	    bp->parent_offset > parent->length - sizeof(binder_uintptr_t)) {
		/* No space for a pointer here! */
		binder_user_error("%d:%d got transaction with invalid parent offset\n",
				  proc->pid, thread->pid);
		return -EINVAL;
	}
	parent_buffer = (u8 *)(parent->buffer -
2380 2381
			binder_alloc_get_user_buffer_offset(
				&target_proc->alloc));
2382 2383 2384 2385 2386
	*(binder_uintptr_t *)(parent_buffer + bp->parent_offset) = bp->buffer;

	return 0;
}

2387 2388
static void binder_transaction(struct binder_proc *proc,
			       struct binder_thread *thread,
2389 2390
			       struct binder_transaction_data *tr, int reply,
			       binder_size_t extra_buffers_size)
2391
{
2392
	int ret;
2393 2394
	struct binder_transaction *t;
	struct binder_work *tcomplete;
2395
	binder_size_t *offp, *off_end, *off_start;
2396
	binder_size_t off_min;
2397
	u8 *sg_bufp, *sg_buf_end;
2398
	struct binder_proc *target_proc = NULL;
2399 2400 2401 2402 2403 2404
	struct binder_thread *target_thread = NULL;
	struct binder_node *target_node = NULL;
	struct list_head *target_list;
	wait_queue_head_t *target_wait;
	struct binder_transaction *in_reply_to = NULL;
	struct binder_transaction_log_entry *e;
2405 2406 2407
	uint32_t return_error = 0;
	uint32_t return_error_param = 0;
	uint32_t return_error_line = 0;
2408 2409
	struct binder_buffer_object *last_fixup_obj = NULL;
	binder_size_t last_fixup_min_off = 0;
2410
	struct binder_context *context = proc->context;
2411
	int t_debug_id = atomic_inc_return(&binder_last_id);
2412 2413

	e = binder_transaction_log_add(&binder_transaction_log);
2414
	e->debug_id = t_debug_id;
2415 2416 2417 2418 2419 2420
	e->call_type = reply ? 2 : !!(tr->flags & TF_ONE_WAY);
	e->from_proc = proc->pid;
	e->from_thread = thread->pid;
	e->target_handle = tr->target.handle;
	e->data_size = tr->data_size;
	e->offsets_size = tr->offsets_size;
2421
	e->context_name = proc->context->name;
2422 2423

	if (reply) {
2424
		binder_inner_proc_lock(proc);
2425 2426
		in_reply_to = thread->transaction_stack;
		if (in_reply_to == NULL) {
2427
			binder_inner_proc_unlock(proc);
2428
			binder_user_error("%d:%d got reply transaction with no transaction stack\n",
2429 2430
					  proc->pid, thread->pid);
			return_error = BR_FAILED_REPLY;
2431 2432
			return_error_param = -EPROTO;
			return_error_line = __LINE__;
2433 2434 2435
			goto err_empty_call_stack;
		}
		if (in_reply_to->to_thread != thread) {
2436
			spin_lock(&in_reply_to->lock);
2437
			binder_user_error("%d:%d got reply transaction with bad transaction stack, transaction %d has target %d:%d\n",
2438 2439 2440 2441 2442
				proc->pid, thread->pid, in_reply_to->debug_id,
				in_reply_to->to_proc ?
				in_reply_to->to_proc->pid : 0,
				in_reply_to->to_thread ?
				in_reply_to->to_thread->pid : 0);
2443
			spin_unlock(&in_reply_to->lock);
2444
			binder_inner_proc_unlock(proc);
2445
			return_error = BR_FAILED_REPLY;
2446 2447
			return_error_param = -EPROTO;
			return_error_line = __LINE__;
2448 2449 2450 2451
			in_reply_to = NULL;
			goto err_bad_call_stack;
		}
		thread->transaction_stack = in_reply_to->to_parent;
2452 2453 2454
		binder_inner_proc_unlock(proc);
		binder_set_nice(in_reply_to->saved_priority);
		target_thread = binder_get_txn_from_and_acq_inner(in_reply_to);
2455 2456
		if (target_thread == NULL) {
			return_error = BR_DEAD_REPLY;
2457
			return_error_line = __LINE__;
2458 2459 2460
			goto err_dead_binder;
		}
		if (target_thread->transaction_stack != in_reply_to) {
2461
			binder_user_error("%d:%d got reply transaction with bad target transaction stack %d, expected %d\n",
2462 2463 2464 2465
				proc->pid, thread->pid,
				target_thread->transaction_stack ?
				target_thread->transaction_stack->debug_id : 0,
				in_reply_to->debug_id);
2466
			binder_inner_proc_unlock(target_thread->proc);
2467
			return_error = BR_FAILED_REPLY;
2468 2469
			return_error_param = -EPROTO;
			return_error_line = __LINE__;
2470 2471 2472 2473 2474
			in_reply_to = NULL;
			target_thread = NULL;
			goto err_dead_binder;
		}
		target_proc = target_thread->proc;
2475
		target_proc->tmp_ref++;
2476
		binder_inner_proc_unlock(target_thread->proc);
2477 2478 2479
	} else {
		if (tr->target.handle) {
			struct binder_ref *ref;
2480

2481 2482 2483 2484 2485 2486 2487
			/*
			 * There must already be a strong ref
			 * on this node. If so, do a strong
			 * increment on the node to ensure it
			 * stays alive until the transaction is
			 * done.
			 */
2488 2489 2490
			binder_proc_lock(proc);
			ref = binder_get_ref_olocked(proc, tr->target.handle,
						     true);
2491 2492 2493 2494
			if (ref) {
				binder_inc_node(ref->node, 1, 0, NULL);
				target_node = ref->node;
			}
2495
			binder_proc_unlock(proc);
2496
			if (target_node == NULL) {
2497
				binder_user_error("%d:%d got transaction to invalid handle\n",
2498 2499
					proc->pid, thread->pid);
				return_error = BR_FAILED_REPLY;
2500 2501
				return_error_param = -EINVAL;
				return_error_line = __LINE__;
2502 2503 2504
				goto err_invalid_target_handle;
			}
		} else {
2505
			mutex_lock(&context->context_mgr_node_lock);
2506
			target_node = context->binder_context_mgr_node;
2507 2508
			if (target_node == NULL) {
				return_error = BR_DEAD_REPLY;
2509
				mutex_unlock(&context->context_mgr_node_lock);
2510
				return_error_line = __LINE__;
2511 2512
				goto err_no_context_mgr_node;
			}
2513
			binder_inc_node(target_node, 1, 0, NULL);
2514
			mutex_unlock(&context->context_mgr_node_lock);
2515 2516
		}
		e->to_node = target_node->debug_id;
2517
		binder_node_lock(target_node);
2518 2519
		target_proc = target_node->proc;
		if (target_proc == NULL) {
2520
			binder_node_unlock(target_node);
2521
			return_error = BR_DEAD_REPLY;
2522
			return_error_line = __LINE__;
2523 2524
			goto err_dead_binder;
		}
2525
		binder_inner_proc_lock(target_proc);
2526
		target_proc->tmp_ref++;
2527
		binder_inner_proc_unlock(target_proc);
2528
		binder_node_unlock(target_node);
2529 2530 2531
		if (security_binder_transaction(proc->tsk,
						target_proc->tsk) < 0) {
			return_error = BR_FAILED_REPLY;
2532 2533
			return_error_param = -EPERM;
			return_error_line = __LINE__;
2534 2535
			goto err_invalid_target_handle;
		}
2536
		binder_inner_proc_lock(proc);
2537 2538
		if (!(tr->flags & TF_ONE_WAY) && thread->transaction_stack) {
			struct binder_transaction *tmp;
2539

2540 2541
			tmp = thread->transaction_stack;
			if (tmp->to_thread != thread) {
2542
				spin_lock(&tmp->lock);
2543
				binder_user_error("%d:%d got new transaction with bad transaction stack, transaction %d has target %d:%d\n",
2544 2545 2546 2547
					proc->pid, thread->pid, tmp->debug_id,
					tmp->to_proc ? tmp->to_proc->pid : 0,
					tmp->to_thread ?
					tmp->to_thread->pid : 0);
2548
				spin_unlock(&tmp->lock);
2549
				binder_inner_proc_unlock(proc);
2550
				return_error = BR_FAILED_REPLY;
2551 2552
				return_error_param = -EPROTO;
				return_error_line = __LINE__;
2553 2554 2555
				goto err_bad_call_stack;
			}
			while (tmp) {
2556 2557 2558 2559 2560 2561 2562 2563 2564 2565 2566
				struct binder_thread *from;

				spin_lock(&tmp->lock);
				from = tmp->from;
				if (from && from->proc == target_proc) {
					atomic_inc(&from->tmp_ref);
					target_thread = from;
					spin_unlock(&tmp->lock);
					break;
				}
				spin_unlock(&tmp->lock);
2567 2568 2569
				tmp = tmp->from_parent;
			}
		}
2570
		binder_inner_proc_unlock(proc);
2571 2572 2573 2574 2575 2576 2577 2578 2579 2580 2581 2582 2583 2584 2585
	}
	if (target_thread) {
		e->to_thread = target_thread->pid;
		target_list = &target_thread->todo;
		target_wait = &target_thread->wait;
	} else {
		target_list = &target_proc->todo;
		target_wait = &target_proc->wait;
	}
	e->to_proc = target_proc->pid;

	/* TODO: reuse incoming transaction for reply */
	t = kzalloc(sizeof(*t), GFP_KERNEL);
	if (t == NULL) {
		return_error = BR_FAILED_REPLY;
2586 2587
		return_error_param = -ENOMEM;
		return_error_line = __LINE__;
2588 2589 2590
		goto err_alloc_t_failed;
	}
	binder_stats_created(BINDER_STAT_TRANSACTION);
2591
	spin_lock_init(&t->lock);
2592 2593 2594 2595

	tcomplete = kzalloc(sizeof(*tcomplete), GFP_KERNEL);
	if (tcomplete == NULL) {
		return_error = BR_FAILED_REPLY;
2596 2597
		return_error_param = -ENOMEM;
		return_error_line = __LINE__;
2598 2599 2600 2601
		goto err_alloc_tcomplete_failed;
	}
	binder_stats_created(BINDER_STAT_TRANSACTION_COMPLETE);

2602
	t->debug_id = t_debug_id;
2603 2604 2605

	if (reply)
		binder_debug(BINDER_DEBUG_TRANSACTION,
2606
			     "%d:%d BC_REPLY %d -> %d:%d, data %016llx-%016llx size %lld-%lld-%lld\n",
2607 2608
			     proc->pid, thread->pid, t->debug_id,
			     target_proc->pid, target_thread->pid,
2609 2610
			     (u64)tr->data.ptr.buffer,
			     (u64)tr->data.ptr.offsets,
2611 2612
			     (u64)tr->data_size, (u64)tr->offsets_size,
			     (u64)extra_buffers_size);
2613 2614
	else
		binder_debug(BINDER_DEBUG_TRANSACTION,
2615
			     "%d:%d BC_TRANSACTION %d -> %d - node %d, data %016llx-%016llx size %lld-%lld-%lld\n",
2616 2617
			     proc->pid, thread->pid, t->debug_id,
			     target_proc->pid, target_node->debug_id,
2618 2619
			     (u64)tr->data.ptr.buffer,
			     (u64)tr->data.ptr.offsets,
2620 2621
			     (u64)tr->data_size, (u64)tr->offsets_size,
			     (u64)extra_buffers_size);
2622 2623 2624 2625 2626

	if (!reply && !(tr->flags & TF_ONE_WAY))
		t->from = thread;
	else
		t->from = NULL;
2627
	t->sender_euid = task_euid(proc->tsk);
2628 2629 2630 2631 2632
	t->to_proc = target_proc;
	t->to_thread = target_thread;
	t->code = tr->code;
	t->flags = tr->flags;
	t->priority = task_nice(current);
2633 2634 2635

	trace_binder_transaction(reply, t, target_node);

2636
	t->buffer = binder_alloc_new_buf(&target_proc->alloc, tr->data_size,
2637 2638
		tr->offsets_size, extra_buffers_size,
		!reply && (t->flags & TF_ONE_WAY));
2639 2640 2641 2642 2643 2644 2645 2646 2647
	if (IS_ERR(t->buffer)) {
		/*
		 * -ESRCH indicates VMA cleared. The target is dying.
		 */
		return_error_param = PTR_ERR(t->buffer);
		return_error = return_error_param == -ESRCH ?
			BR_DEAD_REPLY : BR_FAILED_REPLY;
		return_error_line = __LINE__;
		t->buffer = NULL;
2648 2649 2650 2651 2652 2653
		goto err_binder_alloc_buf_failed;
	}
	t->buffer->allow_user_free = 0;
	t->buffer->debug_id = t->debug_id;
	t->buffer->transaction = t;
	t->buffer->target_node = target_node;
2654
	trace_binder_transaction_alloc_buf(t->buffer);
2655 2656 2657
	off_start = (binder_size_t *)(t->buffer->data +
				      ALIGN(tr->data_size, sizeof(void *)));
	offp = off_start;
2658

2659 2660
	if (copy_from_user(t->buffer->data, (const void __user *)(uintptr_t)
			   tr->data.ptr.buffer, tr->data_size)) {
2661 2662
		binder_user_error("%d:%d got transaction with invalid data ptr\n",
				proc->pid, thread->pid);
2663
		return_error = BR_FAILED_REPLY;
2664 2665
		return_error_param = -EFAULT;
		return_error_line = __LINE__;
2666 2667
		goto err_copy_data_failed;
	}
2668 2669
	if (copy_from_user(offp, (const void __user *)(uintptr_t)
			   tr->data.ptr.offsets, tr->offsets_size)) {
2670 2671
		binder_user_error("%d:%d got transaction with invalid offsets ptr\n",
				proc->pid, thread->pid);
2672
		return_error = BR_FAILED_REPLY;
2673 2674
		return_error_param = -EFAULT;
		return_error_line = __LINE__;
2675 2676
		goto err_copy_data_failed;
	}
2677 2678 2679
	if (!IS_ALIGNED(tr->offsets_size, sizeof(binder_size_t))) {
		binder_user_error("%d:%d got transaction with invalid offsets size, %lld\n",
				proc->pid, thread->pid, (u64)tr->offsets_size);
2680
		return_error = BR_FAILED_REPLY;
2681 2682
		return_error_param = -EINVAL;
		return_error_line = __LINE__;
2683 2684
		goto err_bad_offset;
	}
2685 2686 2687 2688 2689
	if (!IS_ALIGNED(extra_buffers_size, sizeof(u64))) {
		binder_user_error("%d:%d got transaction with unaligned buffers size, %lld\n",
				  proc->pid, thread->pid,
				  (u64)extra_buffers_size);
		return_error = BR_FAILED_REPLY;
2690 2691
		return_error_param = -EINVAL;
		return_error_line = __LINE__;
2692 2693 2694 2695 2696
		goto err_bad_offset;
	}
	off_end = (void *)off_start + tr->offsets_size;
	sg_bufp = (u8 *)(PTR_ALIGN(off_end, sizeof(void *)));
	sg_buf_end = sg_bufp + extra_buffers_size;
2697
	off_min = 0;
2698
	for (; offp < off_end; offp++) {
2699 2700
		struct binder_object_header *hdr;
		size_t object_size = binder_validate_object(t->buffer, *offp);
2701

2702 2703
		if (object_size == 0 || *offp < off_min) {
			binder_user_error("%d:%d got transaction with invalid offset (%lld, min %lld max %lld) or object.\n",
2704 2705
					  proc->pid, thread->pid, (u64)*offp,
					  (u64)off_min,
2706
					  (u64)t->buffer->data_size);
2707
			return_error = BR_FAILED_REPLY;
2708 2709
			return_error_param = -EINVAL;
			return_error_line = __LINE__;
2710 2711
			goto err_bad_offset;
		}
2712 2713 2714 2715

		hdr = (struct binder_object_header *)(t->buffer->data + *offp);
		off_min = *offp + object_size;
		switch (hdr->type) {
2716 2717
		case BINDER_TYPE_BINDER:
		case BINDER_TYPE_WEAK_BINDER: {
2718
			struct flat_binder_object *fp;
2719

2720
			fp = to_flat_binder_object(hdr);
2721 2722
			ret = binder_translate_binder(fp, t, thread);
			if (ret < 0) {
2723
				return_error = BR_FAILED_REPLY;
2724 2725
				return_error_param = ret;
				return_error_line = __LINE__;
2726
				goto err_translate_failed;
2727 2728 2729 2730
			}
		} break;
		case BINDER_TYPE_HANDLE:
		case BINDER_TYPE_WEAK_HANDLE: {
2731
			struct flat_binder_object *fp;
2732

2733
			fp = to_flat_binder_object(hdr);
2734 2735
			ret = binder_translate_handle(fp, t, thread);
			if (ret < 0) {
2736
				return_error = BR_FAILED_REPLY;
2737 2738
				return_error_param = ret;
				return_error_line = __LINE__;
2739
				goto err_translate_failed;
2740 2741 2742 2743
			}
		} break;

		case BINDER_TYPE_FD: {
2744
			struct binder_fd_object *fp = to_binder_fd_object(hdr);
2745 2746
			int target_fd = binder_translate_fd(fp->fd, t, thread,
							    in_reply_to);
2747 2748 2749

			if (target_fd < 0) {
				return_error = BR_FAILED_REPLY;
2750 2751
				return_error_param = target_fd;
				return_error_line = __LINE__;
2752
				goto err_translate_failed;
2753
			}
2754 2755
			fp->pad_binder = 0;
			fp->fd = target_fd;
2756
		} break;
2757 2758 2759 2760 2761 2762 2763 2764 2765 2766 2767
		case BINDER_TYPE_FDA: {
			struct binder_fd_array_object *fda =
				to_binder_fd_array_object(hdr);
			struct binder_buffer_object *parent =
				binder_validate_ptr(t->buffer, fda->parent,
						    off_start,
						    offp - off_start);
			if (!parent) {
				binder_user_error("%d:%d got transaction with invalid parent offset or type\n",
						  proc->pid, thread->pid);
				return_error = BR_FAILED_REPLY;
2768 2769
				return_error_param = -EINVAL;
				return_error_line = __LINE__;
2770 2771 2772 2773 2774 2775 2776 2777 2778
				goto err_bad_parent;
			}
			if (!binder_validate_fixup(t->buffer, off_start,
						   parent, fda->parent_offset,
						   last_fixup_obj,
						   last_fixup_min_off)) {
				binder_user_error("%d:%d got transaction with out-of-order buffer fixup\n",
						  proc->pid, thread->pid);
				return_error = BR_FAILED_REPLY;
2779 2780
				return_error_param = -EINVAL;
				return_error_line = __LINE__;
2781 2782 2783 2784 2785 2786
				goto err_bad_parent;
			}
			ret = binder_translate_fd_array(fda, parent, t, thread,
							in_reply_to);
			if (ret < 0) {
				return_error = BR_FAILED_REPLY;
2787 2788
				return_error_param = ret;
				return_error_line = __LINE__;
2789 2790 2791 2792 2793 2794
				goto err_translate_failed;
			}
			last_fixup_obj = parent;
			last_fixup_min_off =
				fda->parent_offset + sizeof(u32) * fda->num_fds;
		} break;
2795 2796 2797 2798 2799 2800 2801 2802 2803
		case BINDER_TYPE_PTR: {
			struct binder_buffer_object *bp =
				to_binder_buffer_object(hdr);
			size_t buf_left = sg_buf_end - sg_bufp;

			if (bp->length > buf_left) {
				binder_user_error("%d:%d got transaction with too large buffer\n",
						  proc->pid, thread->pid);
				return_error = BR_FAILED_REPLY;
2804 2805
				return_error_param = -EINVAL;
				return_error_line = __LINE__;
2806 2807 2808 2809 2810 2811 2812
				goto err_bad_offset;
			}
			if (copy_from_user(sg_bufp,
					   (const void __user *)(uintptr_t)
					   bp->buffer, bp->length)) {
				binder_user_error("%d:%d got transaction with invalid offsets ptr\n",
						  proc->pid, thread->pid);
2813
				return_error_param = -EFAULT;
2814
				return_error = BR_FAILED_REPLY;
2815
				return_error_line = __LINE__;
2816 2817 2818 2819
				goto err_copy_data_failed;
			}
			/* Fixup buffer pointer to target proc address space */
			bp->buffer = (uintptr_t)sg_bufp +
2820 2821
				binder_alloc_get_user_buffer_offset(
						&target_proc->alloc);
2822 2823 2824 2825 2826 2827 2828 2829
			sg_bufp += ALIGN(bp->length, sizeof(u64));

			ret = binder_fixup_parent(t, thread, bp, off_start,
						  offp - off_start,
						  last_fixup_obj,
						  last_fixup_min_off);
			if (ret < 0) {
				return_error = BR_FAILED_REPLY;
2830 2831
				return_error_param = ret;
				return_error_line = __LINE__;
2832 2833 2834 2835 2836
				goto err_translate_failed;
			}
			last_fixup_obj = bp;
			last_fixup_min_off = 0;
		} break;
2837
		default:
2838
			binder_user_error("%d:%d got transaction with invalid object type, %x\n",
2839
				proc->pid, thread->pid, hdr->type);
2840
			return_error = BR_FAILED_REPLY;
2841 2842
			return_error_param = -EINVAL;
			return_error_line = __LINE__;
2843 2844 2845
			goto err_bad_object_type;
		}
	}
2846
	tcomplete->type = BINDER_WORK_TRANSACTION_COMPLETE;
2847
	binder_enqueue_work(proc, tcomplete, &thread->todo);
2848
	t->work.type = BINDER_WORK_TRANSACTION;
2849

2850
	if (reply) {
2851 2852 2853
		binder_inner_proc_lock(target_proc);
		if (target_thread->is_dead) {
			binder_inner_proc_unlock(target_proc);
2854
			goto err_dead_proc_or_thread;
2855
		}
2856
		BUG_ON(t->buffer->async_transaction != 0);
2857 2858 2859
		binder_pop_transaction_ilocked(target_thread, in_reply_to);
		binder_enqueue_work_ilocked(&t->work, target_list);
		binder_inner_proc_unlock(target_proc);
2860
		binder_free_transaction(in_reply_to);
2861 2862
	} else if (!(t->flags & TF_ONE_WAY)) {
		BUG_ON(t->buffer->async_transaction != 0);
2863
		binder_inner_proc_lock(proc);
2864 2865 2866
		t->need_reply = 1;
		t->from_parent = thread->transaction_stack;
		thread->transaction_stack = t;
2867 2868
		binder_inner_proc_unlock(proc);
		binder_inner_proc_lock(target_proc);
2869 2870
		if (target_proc->is_dead ||
				(target_thread && target_thread->is_dead)) {
2871 2872 2873 2874
			binder_inner_proc_unlock(target_proc);
			binder_inner_proc_lock(proc);
			binder_pop_transaction_ilocked(thread, t);
			binder_inner_proc_unlock(proc);
2875 2876
			goto err_dead_proc_or_thread;
		}
2877 2878
		binder_enqueue_work_ilocked(&t->work, target_list);
		binder_inner_proc_unlock(target_proc);
2879 2880 2881
	} else {
		BUG_ON(target_node == NULL);
		BUG_ON(t->buffer->async_transaction != 1);
2882
		binder_node_lock(target_node);
2883 2884 2885 2886 2887
		if (target_node->has_async_transaction) {
			target_list = &target_node->async_todo;
			target_wait = NULL;
		} else
			target_node->has_async_transaction = 1;
2888 2889 2890 2891 2892
		/*
		 * Test/set of has_async_transaction
		 * must be atomic with enqueue on
		 * async_todo
		 */
2893
		binder_inner_proc_lock(target_proc);
2894
		if (target_proc->is_dead ||
2895
				(target_thread && target_thread->is_dead)) {
2896
			binder_inner_proc_unlock(target_proc);
2897
			binder_node_unlock(target_node);
2898
			goto err_dead_proc_or_thread;
2899
		}
2900 2901
		binder_enqueue_work_ilocked(&t->work, target_list);
		binder_inner_proc_unlock(target_proc);
2902
		binder_node_unlock(target_node);
2903
	}
2904
	if (target_wait) {
2905
		if (reply || !(tr->flags & TF_ONE_WAY))
2906 2907 2908 2909
			wake_up_interruptible_sync(target_wait);
		else
			wake_up_interruptible(target_wait);
	}
2910 2911 2912
	if (target_thread)
		binder_thread_dec_tmpref(target_thread);
	binder_proc_dec_tmpref(target_proc);
2913 2914 2915 2916 2917 2918
	/*
	 * write barrier to synchronize with initialization
	 * of log entry
	 */
	smp_wmb();
	WRITE_ONCE(e->debug_id_done, t_debug_id);
2919 2920
	return;

2921 2922 2923
err_dead_proc_or_thread:
	return_error = BR_DEAD_REPLY;
	return_error_line = __LINE__;
2924
err_translate_failed:
2925 2926
err_bad_object_type:
err_bad_offset:
2927
err_bad_parent:
2928
err_copy_data_failed:
2929
	trace_binder_transaction_failed_buffer_release(t->buffer);
2930
	binder_transaction_buffer_release(target_proc, t->buffer, offp);
2931
	target_node = NULL;
2932
	t->buffer->transaction = NULL;
2933
	binder_alloc_free_buf(&target_proc->alloc, t->buffer);
2934 2935 2936 2937 2938 2939 2940 2941 2942 2943 2944 2945
err_binder_alloc_buf_failed:
	kfree(tcomplete);
	binder_stats_deleted(BINDER_STAT_TRANSACTION_COMPLETE);
err_alloc_tcomplete_failed:
	kfree(t);
	binder_stats_deleted(BINDER_STAT_TRANSACTION);
err_alloc_t_failed:
err_bad_call_stack:
err_empty_call_stack:
err_dead_binder:
err_invalid_target_handle:
err_no_context_mgr_node:
2946 2947 2948 2949
	if (target_thread)
		binder_thread_dec_tmpref(target_thread);
	if (target_proc)
		binder_proc_dec_tmpref(target_proc);
2950 2951 2952
	if (target_node)
		binder_dec_node(target_node, 1, 0);

2953
	binder_debug(BINDER_DEBUG_FAILED_TRANSACTION,
2954 2955 2956 2957
		     "%d:%d transaction failed %d/%d, size %lld-%lld line %d\n",
		     proc->pid, thread->pid, return_error, return_error_param,
		     (u64)tr->data_size, (u64)tr->offsets_size,
		     return_error_line);
2958 2959 2960

	{
		struct binder_transaction_log_entry *fe;
2961

2962 2963 2964
		e->return_error = return_error;
		e->return_error_param = return_error_param;
		e->return_error_line = return_error_line;
2965 2966
		fe = binder_transaction_log_add(&binder_transaction_log_failed);
		*fe = *e;
2967 2968 2969 2970 2971 2972 2973
		/*
		 * write barrier to synchronize with initialization
		 * of log entry
		 */
		smp_wmb();
		WRITE_ONCE(e->debug_id_done, t_debug_id);
		WRITE_ONCE(fe->debug_id_done, t_debug_id);
2974 2975
	}

2976
	BUG_ON(thread->return_error.cmd != BR_OK);
2977
	if (in_reply_to) {
2978
		thread->return_error.cmd = BR_TRANSACTION_COMPLETE;
2979 2980 2981
		binder_enqueue_work(thread->proc,
				    &thread->return_error.work,
				    &thread->todo);
2982
		binder_send_failed_reply(in_reply_to, return_error);
2983 2984
	} else {
		thread->return_error.cmd = return_error;
2985 2986 2987
		binder_enqueue_work(thread->proc,
				    &thread->return_error.work,
				    &thread->todo);
2988
	}
2989 2990
}

2991 2992
static int binder_thread_write(struct binder_proc *proc,
			struct binder_thread *thread,
2993 2994
			binder_uintptr_t binder_buffer, size_t size,
			binder_size_t *consumed)
2995 2996
{
	uint32_t cmd;
2997
	struct binder_context *context = proc->context;
2998
	void __user *buffer = (void __user *)(uintptr_t)binder_buffer;
2999 3000 3001
	void __user *ptr = buffer + *consumed;
	void __user *end = buffer + size;

3002
	while (ptr < end && thread->return_error.cmd == BR_OK) {
3003 3004
		int ret;

3005 3006 3007
		if (get_user(cmd, (uint32_t __user *)ptr))
			return -EFAULT;
		ptr += sizeof(uint32_t);
3008
		trace_binder_command(cmd);
3009
		if (_IOC_NR(cmd) < ARRAY_SIZE(binder_stats.bc)) {
3010 3011 3012
			atomic_inc(&binder_stats.bc[_IOC_NR(cmd)]);
			atomic_inc(&proc->stats.bc[_IOC_NR(cmd)]);
			atomic_inc(&thread->stats.bc[_IOC_NR(cmd)]);
3013 3014 3015 3016 3017 3018 3019 3020
		}
		switch (cmd) {
		case BC_INCREFS:
		case BC_ACQUIRE:
		case BC_RELEASE:
		case BC_DECREFS: {
			uint32_t target;
			const char *debug_string;
3021 3022 3023
			bool strong = cmd == BC_ACQUIRE || cmd == BC_RELEASE;
			bool increment = cmd == BC_INCREFS || cmd == BC_ACQUIRE;
			struct binder_ref_data rdata;
3024 3025 3026

			if (get_user(target, (uint32_t __user *)ptr))
				return -EFAULT;
3027

3028
			ptr += sizeof(uint32_t);
3029 3030
			ret = -1;
			if (increment && !target) {
3031 3032 3033
				struct binder_node *ctx_mgr_node;
				mutex_lock(&context->context_mgr_node_lock);
				ctx_mgr_node = context->binder_context_mgr_node;
3034 3035 3036 3037
				if (ctx_mgr_node)
					ret = binder_inc_ref_for_node(
							proc, ctx_mgr_node,
							strong, NULL, &rdata);
3038 3039
				mutex_unlock(&context->context_mgr_node_lock);
			}
3040 3041 3042 3043 3044 3045 3046 3047
			if (ret)
				ret = binder_update_ref_for_handle(
						proc, target, increment, strong,
						&rdata);
			if (!ret && rdata.desc != target) {
				binder_user_error("%d:%d tried to acquire reference to desc %d, got %d instead\n",
					proc->pid, thread->pid,
					target, rdata.desc);
3048 3049 3050 3051 3052 3053 3054 3055 3056 3057 3058 3059 3060 3061
			}
			switch (cmd) {
			case BC_INCREFS:
				debug_string = "IncRefs";
				break;
			case BC_ACQUIRE:
				debug_string = "Acquire";
				break;
			case BC_RELEASE:
				debug_string = "Release";
				break;
			case BC_DECREFS:
			default:
				debug_string = "DecRefs";
3062 3063 3064 3065 3066 3067
				break;
			}
			if (ret) {
				binder_user_error("%d:%d %s %d refcount change on invalid ref %d ret %d\n",
					proc->pid, thread->pid, debug_string,
					strong, target, ret);
3068 3069 3070
				break;
			}
			binder_debug(BINDER_DEBUG_USER_REFS,
3071 3072 3073 3074
				     "%d:%d %s ref %d desc %d s %d w %d\n",
				     proc->pid, thread->pid, debug_string,
				     rdata.debug_id, rdata.desc, rdata.strong,
				     rdata.weak);
3075 3076 3077 3078
			break;
		}
		case BC_INCREFS_DONE:
		case BC_ACQUIRE_DONE: {
3079 3080
			binder_uintptr_t node_ptr;
			binder_uintptr_t cookie;
3081
			struct binder_node *node;
3082
			bool free_node;
3083

3084
			if (get_user(node_ptr, (binder_uintptr_t __user *)ptr))
3085
				return -EFAULT;
3086 3087
			ptr += sizeof(binder_uintptr_t);
			if (get_user(cookie, (binder_uintptr_t __user *)ptr))
3088
				return -EFAULT;
3089
			ptr += sizeof(binder_uintptr_t);
3090 3091
			node = binder_get_node(proc, node_ptr);
			if (node == NULL) {
3092
				binder_user_error("%d:%d %s u%016llx no match\n",
3093 3094 3095 3096
					proc->pid, thread->pid,
					cmd == BC_INCREFS_DONE ?
					"BC_INCREFS_DONE" :
					"BC_ACQUIRE_DONE",
3097
					(u64)node_ptr);
3098 3099 3100
				break;
			}
			if (cookie != node->cookie) {
3101
				binder_user_error("%d:%d %s u%016llx node %d cookie mismatch %016llx != %016llx\n",
3102 3103 3104
					proc->pid, thread->pid,
					cmd == BC_INCREFS_DONE ?
					"BC_INCREFS_DONE" : "BC_ACQUIRE_DONE",
3105 3106
					(u64)node_ptr, node->debug_id,
					(u64)cookie, (u64)node->cookie);
3107
				binder_put_node(node);
3108 3109
				break;
			}
3110
			binder_node_inner_lock(node);
3111 3112
			if (cmd == BC_ACQUIRE_DONE) {
				if (node->pending_strong_ref == 0) {
3113
					binder_user_error("%d:%d BC_ACQUIRE_DONE node %d has no pending acquire request\n",
3114 3115
						proc->pid, thread->pid,
						node->debug_id);
3116
					binder_node_inner_unlock(node);
3117
					binder_put_node(node);
3118 3119 3120 3121 3122
					break;
				}
				node->pending_strong_ref = 0;
			} else {
				if (node->pending_weak_ref == 0) {
3123
					binder_user_error("%d:%d BC_INCREFS_DONE node %d has no pending increfs request\n",
3124 3125
						proc->pid, thread->pid,
						node->debug_id);
3126
					binder_node_inner_unlock(node);
3127
					binder_put_node(node);
3128 3129 3130 3131
					break;
				}
				node->pending_weak_ref = 0;
			}
3132 3133 3134
			free_node = binder_dec_node_nilocked(node,
					cmd == BC_ACQUIRE_DONE, 0);
			WARN_ON(free_node);
3135
			binder_debug(BINDER_DEBUG_USER_REFS,
3136
				     "%d:%d %s node %d ls %d lw %d tr %d\n",
3137 3138
				     proc->pid, thread->pid,
				     cmd == BC_INCREFS_DONE ? "BC_INCREFS_DONE" : "BC_ACQUIRE_DONE",
3139 3140
				     node->debug_id, node->local_strong_refs,
				     node->local_weak_refs, node->tmp_refs);
3141
			binder_node_inner_unlock(node);
3142
			binder_put_node(node);
3143 3144 3145
			break;
		}
		case BC_ATTEMPT_ACQUIRE:
3146
			pr_err("BC_ATTEMPT_ACQUIRE not supported\n");
3147 3148
			return -EINVAL;
		case BC_ACQUIRE_RESULT:
3149
			pr_err("BC_ACQUIRE_RESULT not supported\n");
3150 3151 3152
			return -EINVAL;

		case BC_FREE_BUFFER: {
3153
			binder_uintptr_t data_ptr;
3154 3155
			struct binder_buffer *buffer;

3156
			if (get_user(data_ptr, (binder_uintptr_t __user *)ptr))
3157
				return -EFAULT;
3158
			ptr += sizeof(binder_uintptr_t);
3159

3160 3161
			buffer = binder_alloc_prepare_to_free(&proc->alloc,
							      data_ptr);
3162
			if (buffer == NULL) {
3163 3164
				binder_user_error("%d:%d BC_FREE_BUFFER u%016llx no match\n",
					proc->pid, thread->pid, (u64)data_ptr);
3165 3166 3167
				break;
			}
			if (!buffer->allow_user_free) {
3168 3169
				binder_user_error("%d:%d BC_FREE_BUFFER u%016llx matched unreturned buffer\n",
					proc->pid, thread->pid, (u64)data_ptr);
3170 3171 3172
				break;
			}
			binder_debug(BINDER_DEBUG_FREE_BUFFER,
3173 3174 3175
				     "%d:%d BC_FREE_BUFFER u%016llx found buffer %d for %s transaction\n",
				     proc->pid, thread->pid, (u64)data_ptr,
				     buffer->debug_id,
3176 3177 3178 3179 3180 3181 3182
				     buffer->transaction ? "active" : "finished");

			if (buffer->transaction) {
				buffer->transaction->buffer = NULL;
				buffer->transaction = NULL;
			}
			if (buffer->async_transaction && buffer->target_node) {
3183 3184 3185 3186
				struct binder_node *buf_node;
				struct binder_work *w;

				buf_node = buffer->target_node;
3187
				binder_node_inner_lock(buf_node);
3188 3189 3190 3191 3192 3193
				BUG_ON(!buf_node->has_async_transaction);
				BUG_ON(buf_node->proc != proc);
				w = binder_dequeue_work_head_ilocked(
						&buf_node->async_todo);
				if (!w)
					buf_node->has_async_transaction = 0;
3194
				else
3195 3196
					binder_enqueue_work_ilocked(
							w, &thread->todo);
3197
				binder_node_inner_unlock(buf_node);
3198
			}
3199
			trace_binder_transaction_buffer_release(buffer);
3200
			binder_transaction_buffer_release(proc, buffer, NULL);
3201
			binder_alloc_free_buf(&proc->alloc, buffer);
3202 3203 3204
			break;
		}

3205 3206 3207 3208 3209 3210 3211 3212 3213 3214 3215
		case BC_TRANSACTION_SG:
		case BC_REPLY_SG: {
			struct binder_transaction_data_sg tr;

			if (copy_from_user(&tr, ptr, sizeof(tr)))
				return -EFAULT;
			ptr += sizeof(tr);
			binder_transaction(proc, thread, &tr.transaction_data,
					   cmd == BC_REPLY_SG, tr.buffers_size);
			break;
		}
3216 3217 3218 3219 3220 3221 3222
		case BC_TRANSACTION:
		case BC_REPLY: {
			struct binder_transaction_data tr;

			if (copy_from_user(&tr, ptr, sizeof(tr)))
				return -EFAULT;
			ptr += sizeof(tr);
3223 3224
			binder_transaction(proc, thread, &tr,
					   cmd == BC_REPLY, 0);
3225 3226 3227 3228 3229
			break;
		}

		case BC_REGISTER_LOOPER:
			binder_debug(BINDER_DEBUG_THREADS,
3230
				     "%d:%d BC_REGISTER_LOOPER\n",
3231
				     proc->pid, thread->pid);
3232
			binder_inner_proc_lock(proc);
3233 3234
			if (thread->looper & BINDER_LOOPER_STATE_ENTERED) {
				thread->looper |= BINDER_LOOPER_STATE_INVALID;
3235
				binder_user_error("%d:%d ERROR: BC_REGISTER_LOOPER called after BC_ENTER_LOOPER\n",
3236 3237 3238
					proc->pid, thread->pid);
			} else if (proc->requested_threads == 0) {
				thread->looper |= BINDER_LOOPER_STATE_INVALID;
3239
				binder_user_error("%d:%d ERROR: BC_REGISTER_LOOPER called without request\n",
3240 3241 3242 3243 3244 3245
					proc->pid, thread->pid);
			} else {
				proc->requested_threads--;
				proc->requested_threads_started++;
			}
			thread->looper |= BINDER_LOOPER_STATE_REGISTERED;
3246
			binder_inner_proc_unlock(proc);
3247 3248 3249
			break;
		case BC_ENTER_LOOPER:
			binder_debug(BINDER_DEBUG_THREADS,
3250
				     "%d:%d BC_ENTER_LOOPER\n",
3251 3252 3253
				     proc->pid, thread->pid);
			if (thread->looper & BINDER_LOOPER_STATE_REGISTERED) {
				thread->looper |= BINDER_LOOPER_STATE_INVALID;
3254
				binder_user_error("%d:%d ERROR: BC_ENTER_LOOPER called after BC_REGISTER_LOOPER\n",
3255 3256 3257 3258 3259 3260
					proc->pid, thread->pid);
			}
			thread->looper |= BINDER_LOOPER_STATE_ENTERED;
			break;
		case BC_EXIT_LOOPER:
			binder_debug(BINDER_DEBUG_THREADS,
3261
				     "%d:%d BC_EXIT_LOOPER\n",
3262 3263 3264 3265 3266 3267 3268
				     proc->pid, thread->pid);
			thread->looper |= BINDER_LOOPER_STATE_EXITED;
			break;

		case BC_REQUEST_DEATH_NOTIFICATION:
		case BC_CLEAR_DEATH_NOTIFICATION: {
			uint32_t target;
3269
			binder_uintptr_t cookie;
3270
			struct binder_ref *ref;
3271
			struct binder_ref_death *death = NULL;
3272 3273 3274 3275

			if (get_user(target, (uint32_t __user *)ptr))
				return -EFAULT;
			ptr += sizeof(uint32_t);
3276
			if (get_user(cookie, (binder_uintptr_t __user *)ptr))
3277
				return -EFAULT;
3278
			ptr += sizeof(binder_uintptr_t);
3279 3280 3281 3282 3283 3284 3285 3286 3287 3288 3289 3290 3291 3292 3293 3294 3295 3296 3297 3298 3299 3300 3301
			if (cmd == BC_REQUEST_DEATH_NOTIFICATION) {
				/*
				 * Allocate memory for death notification
				 * before taking lock
				 */
				death = kzalloc(sizeof(*death), GFP_KERNEL);
				if (death == NULL) {
					WARN_ON(thread->return_error.cmd !=
						BR_OK);
					thread->return_error.cmd = BR_ERROR;
					binder_enqueue_work(
						thread->proc,
						&thread->return_error.work,
						&thread->todo);
					binder_debug(
						BINDER_DEBUG_FAILED_TRANSACTION,
						"%d:%d BC_REQUEST_DEATH_NOTIFICATION failed\n",
						proc->pid, thread->pid);
					break;
				}
			}
			binder_proc_lock(proc);
			ref = binder_get_ref_olocked(proc, target, false);
3302
			if (ref == NULL) {
3303
				binder_user_error("%d:%d %s invalid ref %d\n",
3304 3305 3306 3307 3308
					proc->pid, thread->pid,
					cmd == BC_REQUEST_DEATH_NOTIFICATION ?
					"BC_REQUEST_DEATH_NOTIFICATION" :
					"BC_CLEAR_DEATH_NOTIFICATION",
					target);
3309 3310
				binder_proc_unlock(proc);
				kfree(death);
3311 3312 3313 3314
				break;
			}

			binder_debug(BINDER_DEBUG_DEATH_NOTIFICATION,
3315
				     "%d:%d %s %016llx ref %d desc %d s %d w %d for node %d\n",
3316 3317 3318 3319
				     proc->pid, thread->pid,
				     cmd == BC_REQUEST_DEATH_NOTIFICATION ?
				     "BC_REQUEST_DEATH_NOTIFICATION" :
				     "BC_CLEAR_DEATH_NOTIFICATION",
3320 3321 3322
				     (u64)cookie, ref->data.debug_id,
				     ref->data.desc, ref->data.strong,
				     ref->data.weak, ref->node->debug_id);
3323

3324
			binder_node_lock(ref->node);
3325 3326
			if (cmd == BC_REQUEST_DEATH_NOTIFICATION) {
				if (ref->death) {
3327
					binder_user_error("%d:%d BC_REQUEST_DEATH_NOTIFICATION death notification already set\n",
3328
						proc->pid, thread->pid);
3329
					binder_node_unlock(ref->node);
3330 3331
					binder_proc_unlock(proc);
					kfree(death);
3332 3333 3334 3335 3336 3337 3338 3339
					break;
				}
				binder_stats_created(BINDER_STAT_DEATH);
				INIT_LIST_HEAD(&death->work.entry);
				death->cookie = cookie;
				ref->death = death;
				if (ref->node->proc == NULL) {
					ref->death->work.type = BINDER_WORK_DEAD_BINDER;
3340 3341 3342 3343 3344 3345 3346 3347 3348 3349 3350 3351 3352 3353
					if (thread->looper &
					    (BINDER_LOOPER_STATE_REGISTERED |
					     BINDER_LOOPER_STATE_ENTERED))
						binder_enqueue_work(
							proc,
							&ref->death->work,
							&thread->todo);
					else {
						binder_enqueue_work(
							proc,
							&ref->death->work,
							&proc->todo);
						wake_up_interruptible(
								&proc->wait);
3354 3355 3356 3357
					}
				}
			} else {
				if (ref->death == NULL) {
3358
					binder_user_error("%d:%d BC_CLEAR_DEATH_NOTIFICATION death notification not active\n",
3359
						proc->pid, thread->pid);
3360
					binder_node_unlock(ref->node);
3361
					binder_proc_unlock(proc);
3362 3363 3364 3365
					break;
				}
				death = ref->death;
				if (death->cookie != cookie) {
3366
					binder_user_error("%d:%d BC_CLEAR_DEATH_NOTIFICATION death notification cookie mismatch %016llx != %016llx\n",
3367
						proc->pid, thread->pid,
3368 3369
						(u64)death->cookie,
						(u64)cookie);
3370
					binder_node_unlock(ref->node);
3371
					binder_proc_unlock(proc);
3372 3373 3374
					break;
				}
				ref->death = NULL;
3375
				binder_inner_proc_lock(proc);
3376 3377
				if (list_empty(&death->work.entry)) {
					death->work.type = BINDER_WORK_CLEAR_DEATH_NOTIFICATION;
3378 3379 3380 3381 3382 3383 3384 3385 3386 3387 3388 3389
					if (thread->looper &
					    (BINDER_LOOPER_STATE_REGISTERED |
					     BINDER_LOOPER_STATE_ENTERED))
						binder_enqueue_work_ilocked(
								&death->work,
								&thread->todo);
					else {
						binder_enqueue_work_ilocked(
								&death->work,
								&proc->todo);
						wake_up_interruptible(
								&proc->wait);
3390 3391 3392 3393 3394
					}
				} else {
					BUG_ON(death->work.type != BINDER_WORK_DEAD_BINDER);
					death->work.type = BINDER_WORK_DEAD_BINDER_AND_CLEAR;
				}
3395
				binder_inner_proc_unlock(proc);
3396
			}
3397
			binder_node_unlock(ref->node);
3398
			binder_proc_unlock(proc);
3399 3400 3401
		} break;
		case BC_DEAD_BINDER_DONE: {
			struct binder_work *w;
3402
			binder_uintptr_t cookie;
3403
			struct binder_ref_death *death = NULL;
3404

3405
			if (get_user(cookie, (binder_uintptr_t __user *)ptr))
3406 3407
				return -EFAULT;

3408
			ptr += sizeof(cookie);
3409 3410 3411 3412 3413 3414 3415
			binder_inner_proc_lock(proc);
			list_for_each_entry(w, &proc->delivered_death,
					    entry) {
				struct binder_ref_death *tmp_death =
					container_of(w,
						     struct binder_ref_death,
						     work);
3416

3417 3418 3419 3420 3421 3422
				if (tmp_death->cookie == cookie) {
					death = tmp_death;
					break;
				}
			}
			binder_debug(BINDER_DEBUG_DEAD_BINDER,
3423 3424 3425
				     "%d:%d BC_DEAD_BINDER_DONE %016llx found %p\n",
				     proc->pid, thread->pid, (u64)cookie,
				     death);
3426
			if (death == NULL) {
3427 3428
				binder_user_error("%d:%d BC_DEAD_BINDER_DONE %016llx not found\n",
					proc->pid, thread->pid, (u64)cookie);
3429
				binder_inner_proc_unlock(proc);
3430 3431
				break;
			}
3432
			binder_dequeue_work_ilocked(&death->work);
3433 3434
			if (death->work.type == BINDER_WORK_DEAD_BINDER_AND_CLEAR) {
				death->work.type = BINDER_WORK_CLEAR_DEATH_NOTIFICATION;
3435 3436 3437 3438 3439 3440 3441 3442 3443
				if (thread->looper &
					(BINDER_LOOPER_STATE_REGISTERED |
					 BINDER_LOOPER_STATE_ENTERED))
					binder_enqueue_work_ilocked(
						&death->work, &thread->todo);
				else {
					binder_enqueue_work_ilocked(
							&death->work,
							&proc->todo);
3444 3445 3446
					wake_up_interruptible(&proc->wait);
				}
			}
3447
			binder_inner_proc_unlock(proc);
3448 3449 3450
		} break;

		default:
3451
			pr_err("%d:%d unknown command %d\n",
3452 3453 3454 3455 3456 3457 3458 3459
			       proc->pid, thread->pid, cmd);
			return -EINVAL;
		}
		*consumed = ptr - buffer;
	}
	return 0;
}

3460 3461
static void binder_stat_br(struct binder_proc *proc,
			   struct binder_thread *thread, uint32_t cmd)
3462
{
3463
	trace_binder_return(cmd);
3464
	if (_IOC_NR(cmd) < ARRAY_SIZE(binder_stats.br)) {
3465 3466 3467
		atomic_inc(&binder_stats.br[_IOC_NR(cmd)]);
		atomic_inc(&proc->stats.br[_IOC_NR(cmd)]);
		atomic_inc(&thread->stats.br[_IOC_NR(cmd)]);
3468 3469 3470 3471 3472 3473
	}
}

static int binder_has_proc_work(struct binder_proc *proc,
				struct binder_thread *thread)
{
3474 3475
	return !binder_worklist_empty(proc, &proc->todo) ||
		thread->looper_need_return;
3476 3477 3478 3479
}

static int binder_has_thread_work(struct binder_thread *thread)
{
3480 3481
	return !binder_worklist_empty(thread->proc, &thread->todo) ||
		thread->looper_need_return;
3482 3483
}

3484 3485 3486 3487 3488 3489 3490 3491 3492 3493 3494 3495 3496 3497 3498 3499 3500 3501 3502 3503 3504 3505 3506 3507 3508 3509 3510 3511 3512 3513 3514
static int binder_put_node_cmd(struct binder_proc *proc,
			       struct binder_thread *thread,
			       void __user **ptrp,
			       binder_uintptr_t node_ptr,
			       binder_uintptr_t node_cookie,
			       int node_debug_id,
			       uint32_t cmd, const char *cmd_name)
{
	void __user *ptr = *ptrp;

	if (put_user(cmd, (uint32_t __user *)ptr))
		return -EFAULT;
	ptr += sizeof(uint32_t);

	if (put_user(node_ptr, (binder_uintptr_t __user *)ptr))
		return -EFAULT;
	ptr += sizeof(binder_uintptr_t);

	if (put_user(node_cookie, (binder_uintptr_t __user *)ptr))
		return -EFAULT;
	ptr += sizeof(binder_uintptr_t);

	binder_stat_br(proc, thread, cmd);
	binder_debug(BINDER_DEBUG_USER_REFS, "%d:%d %s %d u%016llx c%016llx\n",
		     proc->pid, thread->pid, cmd_name, node_debug_id,
		     (u64)node_ptr, (u64)node_cookie);

	*ptrp = ptr;
	return 0;
}

3515 3516
static int binder_thread_read(struct binder_proc *proc,
			      struct binder_thread *thread,
3517 3518
			      binder_uintptr_t binder_buffer, size_t size,
			      binder_size_t *consumed, int non_block)
3519
{
3520
	void __user *buffer = (void __user *)(uintptr_t)binder_buffer;
3521 3522 3523 3524 3525 3526 3527 3528 3529 3530 3531 3532 3533
	void __user *ptr = buffer + *consumed;
	void __user *end = buffer + size;

	int ret = 0;
	int wait_for_proc_work;

	if (*consumed == 0) {
		if (put_user(BR_NOOP, (uint32_t __user *)ptr))
			return -EFAULT;
		ptr += sizeof(uint32_t);
	}

retry:
3534
	binder_inner_proc_lock(proc);
3535
	wait_for_proc_work = thread->transaction_stack == NULL &&
3536
		binder_worklist_empty_ilocked(&thread->todo);
3537 3538
	if (wait_for_proc_work)
		proc->ready_threads++;
3539
	binder_inner_proc_unlock(proc);
3540 3541

	thread->looper |= BINDER_LOOPER_STATE_WAITING;
3542 3543 3544

	trace_binder_wait_for_work(wait_for_proc_work,
				   !!thread->transaction_stack,
3545
				   !binder_worklist_empty(proc, &thread->todo));
3546 3547 3548
	if (wait_for_proc_work) {
		if (!(thread->looper & (BINDER_LOOPER_STATE_REGISTERED |
					BINDER_LOOPER_STATE_ENTERED))) {
3549
			binder_user_error("%d:%d ERROR: Thread waiting for process work before calling BC_REGISTER_LOOPER or BC_ENTER_LOOPER (state %x)\n",
3550 3551 3552 3553 3554 3555 3556 3557 3558
				proc->pid, thread->pid, thread->looper);
			wait_event_interruptible(binder_user_error_wait,
						 binder_stop_on_user_error < 2);
		}
		binder_set_nice(proc->default_priority);
		if (non_block) {
			if (!binder_has_proc_work(proc, thread))
				ret = -EAGAIN;
		} else
3559
			ret = wait_event_freezable_exclusive(proc->wait, binder_has_proc_work(proc, thread));
3560 3561 3562 3563 3564
	} else {
		if (non_block) {
			if (!binder_has_thread_work(thread))
				ret = -EAGAIN;
		} else
3565
			ret = wait_event_freezable(thread->wait, binder_has_thread_work(thread));
3566
	}
3567

3568
	binder_inner_proc_lock(proc);
3569 3570
	if (wait_for_proc_work)
		proc->ready_threads--;
3571
	binder_inner_proc_unlock(proc);
3572 3573 3574 3575 3576 3577 3578 3579
	thread->looper &= ~BINDER_LOOPER_STATE_WAITING;

	if (ret)
		return ret;

	while (1) {
		uint32_t cmd;
		struct binder_transaction_data tr;
3580 3581
		struct binder_work *w = NULL;
		struct list_head *list = NULL;
3582
		struct binder_transaction *t = NULL;
3583
		struct binder_thread *t_from;
3584

3585
		binder_inner_proc_lock(proc);
3586 3587 3588 3589 3590 3591 3592 3593
		if (!binder_worklist_empty_ilocked(&thread->todo))
			list = &thread->todo;
		else if (!binder_worklist_empty_ilocked(&proc->todo) &&
			   wait_for_proc_work)
			list = &proc->todo;
		else {
			binder_inner_proc_unlock(proc);

3594
			/* no data added */
3595
			if (ptr - buffer == 4 && !thread->looper_need_return)
3596 3597 3598 3599
				goto retry;
			break;
		}

3600 3601
		if (end - ptr < sizeof(tr) + 4) {
			binder_inner_proc_unlock(proc);
3602
			break;
3603
		}
3604
		w = binder_dequeue_work_head_ilocked(list);
3605 3606 3607

		switch (w->type) {
		case BINDER_WORK_TRANSACTION: {
3608
			binder_inner_proc_unlock(proc);
3609 3610
			t = container_of(w, struct binder_transaction, work);
		} break;
3611 3612 3613 3614 3615
		case BINDER_WORK_RETURN_ERROR: {
			struct binder_error *e = container_of(
					w, struct binder_error, work);

			WARN_ON(e->cmd == BR_OK);
3616
			binder_inner_proc_unlock(proc);
3617 3618 3619 3620 3621
			if (put_user(e->cmd, (uint32_t __user *)ptr))
				return -EFAULT;
			e->cmd = BR_OK;
			ptr += sizeof(uint32_t);

3622
			binder_stat_br(proc, thread, e->cmd);
3623
		} break;
3624
		case BINDER_WORK_TRANSACTION_COMPLETE: {
3625
			binder_inner_proc_unlock(proc);
3626 3627 3628 3629 3630 3631 3632
			cmd = BR_TRANSACTION_COMPLETE;
			if (put_user(cmd, (uint32_t __user *)ptr))
				return -EFAULT;
			ptr += sizeof(uint32_t);

			binder_stat_br(proc, thread, cmd);
			binder_debug(BINDER_DEBUG_TRANSACTION_COMPLETE,
3633
				     "%d:%d BR_TRANSACTION_COMPLETE\n",
3634 3635 3636 3637 3638 3639
				     proc->pid, thread->pid);
			kfree(w);
			binder_stats_deleted(BINDER_STAT_TRANSACTION_COMPLETE);
		} break;
		case BINDER_WORK_NODE: {
			struct binder_node *node = container_of(w, struct binder_node, work);
3640 3641 3642 3643 3644 3645 3646 3647 3648 3649 3650 3651
			int strong, weak;
			binder_uintptr_t node_ptr = node->ptr;
			binder_uintptr_t node_cookie = node->cookie;
			int node_debug_id = node->debug_id;
			int has_weak_ref;
			int has_strong_ref;
			void __user *orig_ptr = ptr;

			BUG_ON(proc != node->proc);
			strong = node->internal_strong_refs ||
					node->local_strong_refs;
			weak = !hlist_empty(&node->refs) ||
3652 3653
					node->local_weak_refs ||
					node->tmp_refs || strong;
3654 3655 3656 3657
			has_strong_ref = node->has_strong_ref;
			has_weak_ref = node->has_weak_ref;

			if (weak && !has_weak_ref) {
3658 3659 3660
				node->has_weak_ref = 1;
				node->pending_weak_ref = 1;
				node->local_weak_refs++;
3661 3662
			}
			if (strong && !has_strong_ref) {
3663 3664 3665
				node->has_strong_ref = 1;
				node->pending_strong_ref = 1;
				node->local_strong_refs++;
3666 3667
			}
			if (!strong && has_strong_ref)
3668
				node->has_strong_ref = 0;
3669
			if (!weak && has_weak_ref)
3670
				node->has_weak_ref = 0;
3671 3672 3673 3674 3675 3676 3677 3678
			if (!weak && !strong) {
				binder_debug(BINDER_DEBUG_INTERNAL_REFS,
					     "%d:%d node %d u%016llx c%016llx deleted\n",
					     proc->pid, thread->pid,
					     node_debug_id,
					     (u64)node_ptr,
					     (u64)node_cookie);
				rb_erase(&node->rb_node, &proc->nodes);
3679
				binder_inner_proc_unlock(proc);
3680 3681 3682 3683 3684 3685 3686 3687 3688 3689 3690
				binder_node_lock(node);
				/*
				 * Acquire the node lock before freeing the
				 * node to serialize with other threads that
				 * may have been holding the node lock while
				 * decrementing this node (avoids race where
				 * this thread frees while the other thread
				 * is unlocking the node after the final
				 * decrement)
				 */
				binder_node_unlock(node);
3691 3692 3693 3694
				binder_free_node(node);
			} else
				binder_inner_proc_unlock(proc);

3695 3696 3697 3698 3699 3700 3701 3702 3703 3704 3705 3706 3707 3708 3709 3710 3711 3712 3713 3714 3715 3716 3717 3718 3719 3720 3721 3722 3723
			if (weak && !has_weak_ref)
				ret = binder_put_node_cmd(
						proc, thread, &ptr, node_ptr,
						node_cookie, node_debug_id,
						BR_INCREFS, "BR_INCREFS");
			if (!ret && strong && !has_strong_ref)
				ret = binder_put_node_cmd(
						proc, thread, &ptr, node_ptr,
						node_cookie, node_debug_id,
						BR_ACQUIRE, "BR_ACQUIRE");
			if (!ret && !strong && has_strong_ref)
				ret = binder_put_node_cmd(
						proc, thread, &ptr, node_ptr,
						node_cookie, node_debug_id,
						BR_RELEASE, "BR_RELEASE");
			if (!ret && !weak && has_weak_ref)
				ret = binder_put_node_cmd(
						proc, thread, &ptr, node_ptr,
						node_cookie, node_debug_id,
						BR_DECREFS, "BR_DECREFS");
			if (orig_ptr == ptr)
				binder_debug(BINDER_DEBUG_INTERNAL_REFS,
					     "%d:%d node %d u%016llx c%016llx state unchanged\n",
					     proc->pid, thread->pid,
					     node_debug_id,
					     (u64)node_ptr,
					     (u64)node_cookie);
			if (ret)
				return ret;
3724 3725 3726 3727 3728 3729
		} break;
		case BINDER_WORK_DEAD_BINDER:
		case BINDER_WORK_DEAD_BINDER_AND_CLEAR:
		case BINDER_WORK_CLEAR_DEATH_NOTIFICATION: {
			struct binder_ref_death *death;
			uint32_t cmd;
3730
			binder_uintptr_t cookie;
3731 3732 3733 3734 3735 3736

			death = container_of(w, struct binder_ref_death, work);
			if (w->type == BINDER_WORK_CLEAR_DEATH_NOTIFICATION)
				cmd = BR_CLEAR_DEATH_NOTIFICATION_DONE;
			else
				cmd = BR_DEAD_BINDER;
3737 3738
			cookie = death->cookie;

3739
			binder_debug(BINDER_DEBUG_DEATH_NOTIFICATION,
3740
				     "%d:%d %s %016llx\n",
3741 3742 3743 3744
				      proc->pid, thread->pid,
				      cmd == BR_DEAD_BINDER ?
				      "BR_DEAD_BINDER" :
				      "BR_CLEAR_DEATH_NOTIFICATION_DONE",
3745
				      (u64)cookie);
3746
			if (w->type == BINDER_WORK_CLEAR_DEATH_NOTIFICATION) {
3747
				binder_inner_proc_unlock(proc);
3748 3749
				kfree(death);
				binder_stats_deleted(BINDER_STAT_DEATH);
3750
			} else {
3751 3752
				binder_enqueue_work_ilocked(
						w, &proc->delivered_death);
3753 3754
				binder_inner_proc_unlock(proc);
			}
3755 3756 3757 3758 3759 3760 3761 3762
			if (put_user(cmd, (uint32_t __user *)ptr))
				return -EFAULT;
			ptr += sizeof(uint32_t);
			if (put_user(cookie,
				     (binder_uintptr_t __user *)ptr))
				return -EFAULT;
			ptr += sizeof(binder_uintptr_t);
			binder_stat_br(proc, thread, cmd);
3763 3764 3765 3766 3767 3768 3769 3770 3771 3772 3773
			if (cmd == BR_DEAD_BINDER)
				goto done; /* DEAD_BINDER notifications can cause transactions */
		} break;
		}

		if (!t)
			continue;

		BUG_ON(t->buffer == NULL);
		if (t->buffer->target_node) {
			struct binder_node *target_node = t->buffer->target_node;
3774

3775 3776 3777 3778 3779 3780 3781 3782 3783 3784 3785
			tr.target.ptr = target_node->ptr;
			tr.cookie =  target_node->cookie;
			t->saved_priority = task_nice(current);
			if (t->priority < target_node->min_priority &&
			    !(t->flags & TF_ONE_WAY))
				binder_set_nice(t->priority);
			else if (!(t->flags & TF_ONE_WAY) ||
				 t->saved_priority > target_node->min_priority)
				binder_set_nice(target_node->min_priority);
			cmd = BR_TRANSACTION;
		} else {
3786 3787
			tr.target.ptr = 0;
			tr.cookie = 0;
3788 3789 3790 3791
			cmd = BR_REPLY;
		}
		tr.code = t->code;
		tr.flags = t->flags;
3792
		tr.sender_euid = from_kuid(current_user_ns(), t->sender_euid);
3793

3794 3795 3796
		t_from = binder_get_txn_from(t);
		if (t_from) {
			struct task_struct *sender = t_from->proc->tsk;
3797

3798
			tr.sender_pid = task_tgid_nr_ns(sender,
3799
							task_active_pid_ns(current));
3800 3801 3802 3803 3804 3805
		} else {
			tr.sender_pid = 0;
		}

		tr.data_size = t->buffer->data_size;
		tr.offsets_size = t->buffer->offsets_size;
3806 3807 3808
		tr.data.ptr.buffer = (binder_uintptr_t)
			((uintptr_t)t->buffer->data +
			binder_alloc_get_user_buffer_offset(&proc->alloc));
3809 3810 3811 3812
		tr.data.ptr.offsets = tr.data.ptr.buffer +
					ALIGN(t->buffer->data_size,
					    sizeof(void *));

3813 3814 3815
		if (put_user(cmd, (uint32_t __user *)ptr)) {
			if (t_from)
				binder_thread_dec_tmpref(t_from);
3816
			return -EFAULT;
3817
		}
3818
		ptr += sizeof(uint32_t);
3819 3820 3821
		if (copy_to_user(ptr, &tr, sizeof(tr))) {
			if (t_from)
				binder_thread_dec_tmpref(t_from);
3822
			return -EFAULT;
3823
		}
3824 3825
		ptr += sizeof(tr);

3826
		trace_binder_transaction_received(t);
3827 3828
		binder_stat_br(proc, thread, cmd);
		binder_debug(BINDER_DEBUG_TRANSACTION,
3829
			     "%d:%d %s %d %d:%d, cmd %d size %zd-%zd ptr %016llx-%016llx\n",
3830 3831 3832
			     proc->pid, thread->pid,
			     (cmd == BR_TRANSACTION) ? "BR_TRANSACTION" :
			     "BR_REPLY",
3833 3834
			     t->debug_id, t_from ? t_from->proc->pid : 0,
			     t_from ? t_from->pid : 0, cmd,
3835
			     t->buffer->data_size, t->buffer->offsets_size,
3836
			     (u64)tr.data.ptr.buffer, (u64)tr.data.ptr.offsets);
3837

3838 3839
		if (t_from)
			binder_thread_dec_tmpref(t_from);
3840 3841
		t->buffer->allow_user_free = 1;
		if (cmd == BR_TRANSACTION && !(t->flags & TF_ONE_WAY)) {
3842
			binder_inner_proc_lock(thread->proc);
3843 3844 3845
			t->to_parent = thread->transaction_stack;
			t->to_thread = thread;
			thread->transaction_stack = t;
3846
			binder_inner_proc_unlock(thread->proc);
3847
		} else {
3848
			binder_free_transaction(t);
3849 3850 3851 3852 3853 3854 3855
		}
		break;
	}

done:

	*consumed = ptr - buffer;
3856
	binder_inner_proc_lock(proc);
3857 3858 3859 3860 3861 3862
	if (proc->requested_threads + proc->ready_threads == 0 &&
	    proc->requested_threads_started < proc->max_threads &&
	    (thread->looper & (BINDER_LOOPER_STATE_REGISTERED |
	     BINDER_LOOPER_STATE_ENTERED)) /* the user-space code fails to */
	     /*spawn a new thread if we leave this out */) {
		proc->requested_threads++;
3863
		binder_inner_proc_unlock(proc);
3864
		binder_debug(BINDER_DEBUG_THREADS,
3865
			     "%d:%d BR_SPAWN_LOOPER\n",
3866 3867 3868
			     proc->pid, thread->pid);
		if (put_user(BR_SPAWN_LOOPER, (uint32_t __user *)buffer))
			return -EFAULT;
3869
		binder_stat_br(proc, thread, BR_SPAWN_LOOPER);
3870 3871
	} else
		binder_inner_proc_unlock(proc);
3872 3873 3874
	return 0;
}

3875 3876
static void binder_release_work(struct binder_proc *proc,
				struct list_head *list)
3877 3878
{
	struct binder_work *w;
3879

3880 3881 3882 3883 3884
	while (1) {
		w = binder_dequeue_work_head(proc, list);
		if (!w)
			return;

3885 3886 3887 3888 3889
		switch (w->type) {
		case BINDER_WORK_TRANSACTION: {
			struct binder_transaction *t;

			t = container_of(w, struct binder_transaction, work);
3890 3891
			if (t->buffer->target_node &&
			    !(t->flags & TF_ONE_WAY)) {
3892
				binder_send_failed_reply(t, BR_DEAD_REPLY);
3893 3894
			} else {
				binder_debug(BINDER_DEBUG_DEAD_TRANSACTION,
3895
					"undelivered transaction %d\n",
3896
					t->debug_id);
3897
				binder_free_transaction(t);
3898
			}
3899
		} break;
3900 3901 3902 3903 3904 3905 3906 3907
		case BINDER_WORK_RETURN_ERROR: {
			struct binder_error *e = container_of(
					w, struct binder_error, work);

			binder_debug(BINDER_DEBUG_DEAD_TRANSACTION,
				"undelivered TRANSACTION_ERROR: %u\n",
				e->cmd);
		} break;
3908
		case BINDER_WORK_TRANSACTION_COMPLETE: {
3909
			binder_debug(BINDER_DEBUG_DEAD_TRANSACTION,
3910
				"undelivered TRANSACTION_COMPLETE\n");
3911 3912 3913
			kfree(w);
			binder_stats_deleted(BINDER_STAT_TRANSACTION_COMPLETE);
		} break;
3914 3915 3916 3917 3918 3919
		case BINDER_WORK_DEAD_BINDER_AND_CLEAR:
		case BINDER_WORK_CLEAR_DEATH_NOTIFICATION: {
			struct binder_ref_death *death;

			death = container_of(w, struct binder_ref_death, work);
			binder_debug(BINDER_DEBUG_DEAD_TRANSACTION,
3920 3921
				"undelivered death notification, %016llx\n",
				(u64)death->cookie);
3922 3923 3924
			kfree(death);
			binder_stats_deleted(BINDER_STAT_DEATH);
		} break;
3925
		default:
3926
			pr_err("unexpected work type, %d, not freed\n",
3927
			       w->type);
3928 3929 3930 3931 3932 3933
			break;
		}
	}

}

3934 3935
static struct binder_thread *binder_get_thread_ilocked(
		struct binder_proc *proc, struct binder_thread *new_thread)
3936 3937 3938 3939 3940 3941 3942 3943 3944 3945 3946 3947 3948 3949
{
	struct binder_thread *thread = NULL;
	struct rb_node *parent = NULL;
	struct rb_node **p = &proc->threads.rb_node;

	while (*p) {
		parent = *p;
		thread = rb_entry(parent, struct binder_thread, rb_node);

		if (current->pid < thread->pid)
			p = &(*p)->rb_left;
		else if (current->pid > thread->pid)
			p = &(*p)->rb_right;
		else
3950
			return thread;
3951
	}
3952 3953 3954 3955 3956 3957 3958 3959 3960 3961 3962 3963 3964 3965 3966 3967 3968 3969 3970 3971 3972 3973 3974 3975 3976 3977 3978 3979 3980 3981 3982
	if (!new_thread)
		return NULL;
	thread = new_thread;
	binder_stats_created(BINDER_STAT_THREAD);
	thread->proc = proc;
	thread->pid = current->pid;
	atomic_set(&thread->tmp_ref, 0);
	init_waitqueue_head(&thread->wait);
	INIT_LIST_HEAD(&thread->todo);
	rb_link_node(&thread->rb_node, parent, p);
	rb_insert_color(&thread->rb_node, &proc->threads);
	thread->looper_need_return = true;
	thread->return_error.work.type = BINDER_WORK_RETURN_ERROR;
	thread->return_error.cmd = BR_OK;
	thread->reply_error.work.type = BINDER_WORK_RETURN_ERROR;
	thread->reply_error.cmd = BR_OK;

	return thread;
}

static struct binder_thread *binder_get_thread(struct binder_proc *proc)
{
	struct binder_thread *thread;
	struct binder_thread *new_thread;

	binder_inner_proc_lock(proc);
	thread = binder_get_thread_ilocked(proc, NULL);
	binder_inner_proc_unlock(proc);
	if (!thread) {
		new_thread = kzalloc(sizeof(*thread), GFP_KERNEL);
		if (new_thread == NULL)
3983
			return NULL;
3984 3985 3986 3987 3988
		binder_inner_proc_lock(proc);
		thread = binder_get_thread_ilocked(proc, new_thread);
		binder_inner_proc_unlock(proc);
		if (thread != new_thread)
			kfree(new_thread);
3989 3990 3991 3992
	}
	return thread;
}

3993 3994 3995 3996 3997 3998 3999 4000 4001 4002 4003 4004 4005 4006 4007 4008 4009 4010 4011 4012
static void binder_free_proc(struct binder_proc *proc)
{
	BUG_ON(!list_empty(&proc->todo));
	BUG_ON(!list_empty(&proc->delivered_death));
	binder_alloc_deferred_release(&proc->alloc);
	put_task_struct(proc->tsk);
	binder_stats_deleted(BINDER_STAT_PROC);
	kfree(proc);
}

static void binder_free_thread(struct binder_thread *thread)
{
	BUG_ON(!list_empty(&thread->todo));
	binder_stats_deleted(BINDER_STAT_THREAD);
	binder_proc_dec_tmpref(thread->proc);
	kfree(thread);
}

static int binder_thread_release(struct binder_proc *proc,
				 struct binder_thread *thread)
4013 4014 4015 4016
{
	struct binder_transaction *t;
	struct binder_transaction *send_reply = NULL;
	int active_transactions = 0;
4017
	struct binder_transaction *last_t = NULL;
4018

4019
	binder_inner_proc_lock(thread->proc);
4020 4021 4022 4023 4024 4025 4026 4027 4028 4029 4030 4031
	/*
	 * take a ref on the proc so it survives
	 * after we remove this thread from proc->threads.
	 * The corresponding dec is when we actually
	 * free the thread in binder_free_thread()
	 */
	proc->tmp_ref++;
	/*
	 * take a ref on this thread to ensure it
	 * survives while we are releasing it
	 */
	atomic_inc(&thread->tmp_ref);
4032 4033
	rb_erase(&thread->rb_node, &proc->threads);
	t = thread->transaction_stack;
4034 4035 4036 4037 4038 4039 4040
	if (t) {
		spin_lock(&t->lock);
		if (t->to_thread == thread)
			send_reply = t;
	}
	thread->is_dead = true;

4041
	while (t) {
4042
		last_t = t;
4043 4044
		active_transactions++;
		binder_debug(BINDER_DEBUG_DEAD_TRANSACTION,
4045 4046
			     "release %d:%d transaction %d %s, still active\n",
			      proc->pid, thread->pid,
4047 4048 4049 4050 4051 4052 4053 4054 4055 4056 4057 4058 4059 4060 4061 4062
			     t->debug_id,
			     (t->to_thread == thread) ? "in" : "out");

		if (t->to_thread == thread) {
			t->to_proc = NULL;
			t->to_thread = NULL;
			if (t->buffer) {
				t->buffer->transaction = NULL;
				t->buffer = NULL;
			}
			t = t->to_parent;
		} else if (t->from == thread) {
			t->from = NULL;
			t = t->from_parent;
		} else
			BUG();
4063 4064 4065
		spin_unlock(&last_t->lock);
		if (t)
			spin_lock(&t->lock);
4066
	}
4067
	binder_inner_proc_unlock(thread->proc);
4068

4069 4070
	if (send_reply)
		binder_send_failed_reply(send_reply, BR_DEAD_REPLY);
4071
	binder_release_work(proc, &thread->todo);
4072
	binder_thread_dec_tmpref(thread);
4073 4074 4075 4076 4077 4078 4079 4080 4081 4082 4083 4084
	return active_transactions;
}

static unsigned int binder_poll(struct file *filp,
				struct poll_table_struct *wait)
{
	struct binder_proc *proc = filp->private_data;
	struct binder_thread *thread = NULL;
	int wait_for_proc_work;

	thread = binder_get_thread(proc);

4085
	binder_inner_proc_lock(thread->proc);
4086
	wait_for_proc_work = thread->transaction_stack == NULL &&
4087 4088
		binder_worklist_empty_ilocked(&thread->todo);
	binder_inner_proc_unlock(thread->proc);
4089

4090 4091 4092 4093 4094 4095 4096 4097 4098 4099 4100 4101 4102 4103 4104 4105
	if (wait_for_proc_work) {
		if (binder_has_proc_work(proc, thread))
			return POLLIN;
		poll_wait(filp, &proc->wait, wait);
		if (binder_has_proc_work(proc, thread))
			return POLLIN;
	} else {
		if (binder_has_thread_work(thread))
			return POLLIN;
		poll_wait(filp, &thread->wait, wait);
		if (binder_has_thread_work(thread))
			return POLLIN;
	}
	return 0;
}

4106 4107 4108 4109 4110 4111 4112 4113 4114 4115 4116 4117 4118 4119 4120 4121 4122 4123 4124 4125 4126 4127 4128 4129 4130 4131 4132 4133 4134 4135 4136 4137 4138 4139 4140 4141 4142 4143 4144 4145 4146 4147 4148
static int binder_ioctl_write_read(struct file *filp,
				unsigned int cmd, unsigned long arg,
				struct binder_thread *thread)
{
	int ret = 0;
	struct binder_proc *proc = filp->private_data;
	unsigned int size = _IOC_SIZE(cmd);
	void __user *ubuf = (void __user *)arg;
	struct binder_write_read bwr;

	if (size != sizeof(struct binder_write_read)) {
		ret = -EINVAL;
		goto out;
	}
	if (copy_from_user(&bwr, ubuf, sizeof(bwr))) {
		ret = -EFAULT;
		goto out;
	}
	binder_debug(BINDER_DEBUG_READ_WRITE,
		     "%d:%d write %lld at %016llx, read %lld at %016llx\n",
		     proc->pid, thread->pid,
		     (u64)bwr.write_size, (u64)bwr.write_buffer,
		     (u64)bwr.read_size, (u64)bwr.read_buffer);

	if (bwr.write_size > 0) {
		ret = binder_thread_write(proc, thread,
					  bwr.write_buffer,
					  bwr.write_size,
					  &bwr.write_consumed);
		trace_binder_write_done(ret);
		if (ret < 0) {
			bwr.read_consumed = 0;
			if (copy_to_user(ubuf, &bwr, sizeof(bwr)))
				ret = -EFAULT;
			goto out;
		}
	}
	if (bwr.read_size > 0) {
		ret = binder_thread_read(proc, thread, bwr.read_buffer,
					 bwr.read_size,
					 &bwr.read_consumed,
					 filp->f_flags & O_NONBLOCK);
		trace_binder_read_done(ret);
4149
		if (!binder_worklist_empty(proc, &proc->todo))
4150 4151 4152 4153 4154 4155 4156 4157 4158 4159 4160 4161 4162 4163 4164 4165 4166 4167 4168 4169 4170 4171 4172 4173
			wake_up_interruptible(&proc->wait);
		if (ret < 0) {
			if (copy_to_user(ubuf, &bwr, sizeof(bwr)))
				ret = -EFAULT;
			goto out;
		}
	}
	binder_debug(BINDER_DEBUG_READ_WRITE,
		     "%d:%d wrote %lld of %lld, read return %lld of %lld\n",
		     proc->pid, thread->pid,
		     (u64)bwr.write_consumed, (u64)bwr.write_size,
		     (u64)bwr.read_consumed, (u64)bwr.read_size);
	if (copy_to_user(ubuf, &bwr, sizeof(bwr))) {
		ret = -EFAULT;
		goto out;
	}
out:
	return ret;
}

static int binder_ioctl_set_ctx_mgr(struct file *filp)
{
	int ret = 0;
	struct binder_proc *proc = filp->private_data;
4174
	struct binder_context *context = proc->context;
4175
	struct binder_node *new_node;
4176 4177
	kuid_t curr_euid = current_euid();

4178
	mutex_lock(&context->context_mgr_node_lock);
4179
	if (context->binder_context_mgr_node) {
4180 4181 4182 4183
		pr_err("BINDER_SET_CONTEXT_MGR already set\n");
		ret = -EBUSY;
		goto out;
	}
4184 4185 4186
	ret = security_binder_set_context_mgr(proc->tsk);
	if (ret < 0)
		goto out;
4187 4188
	if (uid_valid(context->binder_context_mgr_uid)) {
		if (!uid_eq(context->binder_context_mgr_uid, curr_euid)) {
4189 4190 4191
			pr_err("BINDER_SET_CONTEXT_MGR bad uid %d != %d\n",
			       from_kuid(&init_user_ns, curr_euid),
			       from_kuid(&init_user_ns,
4192
					 context->binder_context_mgr_uid));
4193 4194 4195 4196
			ret = -EPERM;
			goto out;
		}
	} else {
4197
		context->binder_context_mgr_uid = curr_euid;
4198
	}
4199
	new_node = binder_new_node(proc, NULL);
4200
	if (!new_node) {
4201 4202 4203
		ret = -ENOMEM;
		goto out;
	}
4204
	binder_node_lock(new_node);
4205 4206 4207 4208 4209
	new_node->local_weak_refs++;
	new_node->local_strong_refs++;
	new_node->has_strong_ref = 1;
	new_node->has_weak_ref = 1;
	context->binder_context_mgr_node = new_node;
4210
	binder_node_unlock(new_node);
4211
	binder_put_node(new_node);
4212
out:
4213
	mutex_unlock(&context->context_mgr_node_lock);
4214 4215 4216
	return ret;
}

4217 4218 4219 4220 4221 4222 4223 4224
static long binder_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
{
	int ret;
	struct binder_proc *proc = filp->private_data;
	struct binder_thread *thread;
	unsigned int size = _IOC_SIZE(cmd);
	void __user *ubuf = (void __user *)arg;

4225 4226
	/*pr_info("binder_ioctl: %d:%d %x %lx\n",
			proc->pid, current->pid, cmd, arg);*/
4227

4228 4229
	binder_selftest_alloc(&proc->alloc);

4230 4231
	trace_binder_ioctl(cmd, arg);

4232 4233
	ret = wait_event_interruptible(binder_user_error_wait, binder_stop_on_user_error < 2);
	if (ret)
4234
		goto err_unlocked;
4235 4236 4237 4238 4239 4240 4241 4242

	thread = binder_get_thread(proc);
	if (thread == NULL) {
		ret = -ENOMEM;
		goto err;
	}

	switch (cmd) {
4243 4244 4245
	case BINDER_WRITE_READ:
		ret = binder_ioctl_write_read(filp, cmd, arg, thread);
		if (ret)
4246 4247
			goto err;
		break;
4248 4249 4250 4251 4252
	case BINDER_SET_MAX_THREADS: {
		int max_threads;

		if (copy_from_user(&max_threads, ubuf,
				   sizeof(max_threads))) {
4253 4254 4255
			ret = -EINVAL;
			goto err;
		}
4256 4257 4258
		binder_inner_proc_lock(proc);
		proc->max_threads = max_threads;
		binder_inner_proc_unlock(proc);
4259
		break;
4260
	}
4261
	case BINDER_SET_CONTEXT_MGR:
4262 4263
		ret = binder_ioctl_set_ctx_mgr(filp);
		if (ret)
4264 4265 4266
			goto err;
		break;
	case BINDER_THREAD_EXIT:
4267
		binder_debug(BINDER_DEBUG_THREADS, "%d:%d exit\n",
4268
			     proc->pid, thread->pid);
4269
		binder_thread_release(proc, thread);
4270 4271
		thread = NULL;
		break;
4272 4273 4274
	case BINDER_VERSION: {
		struct binder_version __user *ver = ubuf;

4275 4276 4277 4278
		if (size != sizeof(struct binder_version)) {
			ret = -EINVAL;
			goto err;
		}
4279 4280
		if (put_user(BINDER_CURRENT_PROTOCOL_VERSION,
			     &ver->protocol_version)) {
4281 4282 4283 4284
			ret = -EINVAL;
			goto err;
		}
		break;
4285
	}
4286 4287 4288 4289 4290 4291 4292
	default:
		ret = -EINVAL;
		goto err;
	}
	ret = 0;
err:
	if (thread)
4293
		thread->looper_need_return = false;
4294 4295
	wait_event_interruptible(binder_user_error_wait, binder_stop_on_user_error < 2);
	if (ret && ret != -ERESTARTSYS)
4296
		pr_info("%d:%d ioctl %x %lx returned %d\n", proc->pid, current->pid, cmd, arg, ret);
4297 4298
err_unlocked:
	trace_binder_ioctl_done(ret);
4299 4300 4301 4302 4303 4304
	return ret;
}

static void binder_vma_open(struct vm_area_struct *vma)
{
	struct binder_proc *proc = vma->vm_private_data;
4305

4306
	binder_debug(BINDER_DEBUG_OPEN_CLOSE,
4307
		     "%d open vm area %lx-%lx (%ld K) vma %lx pagep %lx\n",
4308 4309 4310 4311 4312 4313 4314 4315
		     proc->pid, vma->vm_start, vma->vm_end,
		     (vma->vm_end - vma->vm_start) / SZ_1K, vma->vm_flags,
		     (unsigned long)pgprot_val(vma->vm_page_prot));
}

static void binder_vma_close(struct vm_area_struct *vma)
{
	struct binder_proc *proc = vma->vm_private_data;
4316

4317
	binder_debug(BINDER_DEBUG_OPEN_CLOSE,
4318
		     "%d close vm area %lx-%lx (%ld K) vma %lx pagep %lx\n",
4319 4320 4321
		     proc->pid, vma->vm_start, vma->vm_end,
		     (vma->vm_end - vma->vm_start) / SZ_1K, vma->vm_flags,
		     (unsigned long)pgprot_val(vma->vm_page_prot));
4322
	binder_alloc_vma_close(&proc->alloc);
4323 4324 4325
	binder_defer_work(proc, BINDER_DEFERRED_PUT_FILES);
}

4326
static int binder_vm_fault(struct vm_fault *vmf)
4327 4328 4329 4330
{
	return VM_FAULT_SIGBUS;
}

4331
static const struct vm_operations_struct binder_vm_ops = {
4332 4333
	.open = binder_vma_open,
	.close = binder_vma_close,
4334
	.fault = binder_vm_fault,
4335 4336
};

4337 4338 4339 4340 4341 4342 4343 4344 4345 4346 4347 4348 4349 4350 4351 4352 4353 4354 4355 4356 4357 4358 4359 4360 4361 4362 4363 4364 4365 4366 4367 4368 4369
static int binder_mmap(struct file *filp, struct vm_area_struct *vma)
{
	int ret;
	struct binder_proc *proc = filp->private_data;
	const char *failure_string;

	if (proc->tsk != current->group_leader)
		return -EINVAL;

	if ((vma->vm_end - vma->vm_start) > SZ_4M)
		vma->vm_end = vma->vm_start + SZ_4M;

	binder_debug(BINDER_DEBUG_OPEN_CLOSE,
		     "%s: %d %lx-%lx (%ld K) vma %lx pagep %lx\n",
		     __func__, proc->pid, vma->vm_start, vma->vm_end,
		     (vma->vm_end - vma->vm_start) / SZ_1K, vma->vm_flags,
		     (unsigned long)pgprot_val(vma->vm_page_prot));

	if (vma->vm_flags & FORBIDDEN_MMAP_FLAGS) {
		ret = -EPERM;
		failure_string = "bad vm_flags";
		goto err_bad_arg;
	}
	vma->vm_flags = (vma->vm_flags | VM_DONTCOPY) & ~VM_MAYWRITE;
	vma->vm_ops = &binder_vm_ops;
	vma->vm_private_data = proc;

	ret = binder_alloc_mmap_handler(&proc->alloc, vma);
	if (ret)
		return ret;
	proc->files = get_files_struct(current);
	return 0;

4370
err_bad_arg:
4371
	pr_err("binder_mmap: %d %lx-%lx %s failed %d\n",
4372 4373 4374 4375 4376 4377 4378
	       proc->pid, vma->vm_start, vma->vm_end, failure_string, ret);
	return ret;
}

static int binder_open(struct inode *nodp, struct file *filp)
{
	struct binder_proc *proc;
4379
	struct binder_device *binder_dev;
4380 4381 4382 4383 4384 4385 4386

	binder_debug(BINDER_DEBUG_OPEN_CLOSE, "binder_open: %d:%d\n",
		     current->group_leader->pid, current->pid);

	proc = kzalloc(sizeof(*proc), GFP_KERNEL);
	if (proc == NULL)
		return -ENOMEM;
4387 4388
	spin_lock_init(&proc->inner_lock);
	spin_lock_init(&proc->outer_lock);
4389 4390
	get_task_struct(current->group_leader);
	proc->tsk = current->group_leader;
4391 4392 4393
	INIT_LIST_HEAD(&proc->todo);
	init_waitqueue_head(&proc->wait);
	proc->default_priority = task_nice(current);
4394 4395 4396
	binder_dev = container_of(filp->private_data, struct binder_device,
				  miscdev);
	proc->context = &binder_dev->context;
4397
	binder_alloc_init(&proc->alloc);
4398

4399 4400 4401 4402
	binder_stats_created(BINDER_STAT_PROC);
	proc->pid = current->group_leader->pid;
	INIT_LIST_HEAD(&proc->delivered_death);
	filp->private_data = proc;
4403

4404 4405 4406 4407
	mutex_lock(&binder_procs_lock);
	hlist_add_head(&proc->proc_node, &binder_procs);
	mutex_unlock(&binder_procs_lock);

4408
	if (binder_debugfs_dir_entry_proc) {
4409
		char strbuf[11];
4410

4411
		snprintf(strbuf, sizeof(strbuf), "%u", proc->pid);
4412 4413 4414 4415 4416 4417 4418
		/*
		 * proc debug entries are shared between contexts, so
		 * this will fail if the process tries to open the driver
		 * again with a different context. The priting code will
		 * anyway print all contexts that a given PID has, so this
		 * is not a problem.
		 */
4419
		proc->debugfs_entry = debugfs_create_file(strbuf, S_IRUGO,
4420 4421 4422
			binder_debugfs_dir_entry_proc,
			(void *)(unsigned long)proc->pid,
			&binder_proc_fops);
4423 4424 4425 4426 4427 4428 4429 4430 4431 4432 4433 4434 4435 4436 4437 4438 4439 4440
	}

	return 0;
}

static int binder_flush(struct file *filp, fl_owner_t id)
{
	struct binder_proc *proc = filp->private_data;

	binder_defer_work(proc, BINDER_DEFERRED_FLUSH);

	return 0;
}

static void binder_deferred_flush(struct binder_proc *proc)
{
	struct rb_node *n;
	int wake_count = 0;
4441

4442
	binder_inner_proc_lock(proc);
4443 4444
	for (n = rb_first(&proc->threads); n != NULL; n = rb_next(n)) {
		struct binder_thread *thread = rb_entry(n, struct binder_thread, rb_node);
4445

4446
		thread->looper_need_return = true;
4447 4448 4449 4450 4451
		if (thread->looper & BINDER_LOOPER_STATE_WAITING) {
			wake_up_interruptible(&thread->wait);
			wake_count++;
		}
	}
4452
	binder_inner_proc_unlock(proc);
4453 4454 4455 4456 4457 4458 4459 4460 4461 4462
	wake_up_interruptible_all(&proc->wait);

	binder_debug(BINDER_DEBUG_OPEN_CLOSE,
		     "binder_flush: %d woke %d threads\n", proc->pid,
		     wake_count);
}

static int binder_release(struct inode *nodp, struct file *filp)
{
	struct binder_proc *proc = filp->private_data;
4463

4464
	debugfs_remove(proc->debugfs_entry);
4465 4466 4467 4468 4469
	binder_defer_work(proc, BINDER_DEFERRED_RELEASE);

	return 0;
}

4470 4471 4472 4473
static int binder_node_release(struct binder_node *node, int refs)
{
	struct binder_ref *ref;
	int death = 0;
4474
	struct binder_proc *proc = node->proc;
4475

4476
	binder_release_work(proc, &node->async_todo);
4477

4478
	binder_node_lock(node);
4479
	binder_inner_proc_lock(proc);
4480
	binder_dequeue_work_ilocked(&node->work);
4481 4482 4483 4484 4485
	/*
	 * The caller must have taken a temporary ref on the node,
	 */
	BUG_ON(!node->tmp_refs);
	if (hlist_empty(&node->refs) && node->tmp_refs == 1) {
4486
		binder_inner_proc_unlock(proc);
4487
		binder_node_unlock(node);
4488
		binder_free_node(node);
4489 4490 4491 4492 4493 4494 4495

		return refs;
	}

	node->proc = NULL;
	node->local_strong_refs = 0;
	node->local_weak_refs = 0;
4496
	binder_inner_proc_unlock(proc);
4497 4498

	spin_lock(&binder_dead_nodes_lock);
4499
	hlist_add_head(&node->dead_node, &binder_dead_nodes);
4500
	spin_unlock(&binder_dead_nodes_lock);
4501 4502 4503

	hlist_for_each_entry(ref, &node->refs, node_entry) {
		refs++;
4504 4505 4506 4507 4508 4509 4510 4511 4512
		/*
		 * Need the node lock to synchronize
		 * with new notification requests and the
		 * inner lock to synchronize with queued
		 * death notifications.
		 */
		binder_inner_proc_lock(ref->proc);
		if (!ref->death) {
			binder_inner_proc_unlock(ref->proc);
4513
			continue;
4514
		}
4515 4516 4517

		death++;

4518 4519 4520 4521 4522
		BUG_ON(!list_empty(&ref->death->work.entry));
		ref->death->work.type = BINDER_WORK_DEAD_BINDER;
		binder_enqueue_work_ilocked(&ref->death->work,
					    &ref->proc->todo);
		wake_up_interruptible(&ref->proc->wait);
4523
		binder_inner_proc_unlock(ref->proc);
4524 4525 4526 4527 4528
	}

	binder_debug(BINDER_DEBUG_DEAD_BINDER,
		     "node %d now dead, refs %d, death %d\n",
		     node->debug_id, refs, death);
4529
	binder_node_unlock(node);
4530
	binder_put_node(node);
4531 4532 4533 4534

	return refs;
}

4535 4536
static void binder_deferred_release(struct binder_proc *proc)
{
4537
	struct binder_context *context = proc->context;
4538
	struct rb_node *n;
4539
	int threads, nodes, incoming_refs, outgoing_refs, active_transactions;
4540 4541 4542

	BUG_ON(proc->files);

4543
	mutex_lock(&binder_procs_lock);
4544
	hlist_del(&proc->proc_node);
4545
	mutex_unlock(&binder_procs_lock);
4546

4547
	mutex_lock(&context->context_mgr_node_lock);
4548 4549
	if (context->binder_context_mgr_node &&
	    context->binder_context_mgr_node->proc == proc) {
4550
		binder_debug(BINDER_DEBUG_DEAD_BINDER,
4551 4552
			     "%s: %d context_mgr_node gone\n",
			     __func__, proc->pid);
4553
		context->binder_context_mgr_node = NULL;
4554
	}
4555
	mutex_unlock(&context->context_mgr_node_lock);
4556
	binder_inner_proc_lock(proc);
4557 4558 4559 4560 4561
	/*
	 * Make sure proc stays alive after we
	 * remove all the threads
	 */
	proc->tmp_ref++;
4562

4563
	proc->is_dead = true;
4564 4565 4566
	threads = 0;
	active_transactions = 0;
	while ((n = rb_first(&proc->threads))) {
4567 4568 4569
		struct binder_thread *thread;

		thread = rb_entry(n, struct binder_thread, rb_node);
4570
		binder_inner_proc_unlock(proc);
4571
		threads++;
4572
		active_transactions += binder_thread_release(proc, thread);
4573
		binder_inner_proc_lock(proc);
4574
	}
4575

4576 4577 4578
	nodes = 0;
	incoming_refs = 0;
	while ((n = rb_first(&proc->nodes))) {
4579
		struct binder_node *node;
4580

4581
		node = rb_entry(n, struct binder_node, rb_node);
4582
		nodes++;
4583 4584 4585 4586 4587
		/*
		 * take a temporary ref on the node before
		 * calling binder_node_release() which will either
		 * kfree() the node or call binder_put_node()
		 */
4588
		binder_inc_node_tmpref_ilocked(node);
4589
		rb_erase(&node->rb_node, &proc->nodes);
4590
		binder_inner_proc_unlock(proc);
4591
		incoming_refs = binder_node_release(node, incoming_refs);
4592
		binder_inner_proc_lock(proc);
4593
	}
4594
	binder_inner_proc_unlock(proc);
4595

4596
	outgoing_refs = 0;
4597
	binder_proc_lock(proc);
4598
	while ((n = rb_first(&proc->refs_by_desc))) {
4599 4600 4601
		struct binder_ref *ref;

		ref = rb_entry(n, struct binder_ref, rb_node_desc);
4602
		outgoing_refs++;
4603 4604
		binder_cleanup_ref_olocked(ref);
		binder_proc_unlock(proc);
4605
		binder_free_ref(ref);
4606
		binder_proc_lock(proc);
4607
	}
4608
	binder_proc_unlock(proc);
4609

4610 4611
	binder_release_work(proc, &proc->todo);
	binder_release_work(proc, &proc->delivered_death);
4612 4613

	binder_debug(BINDER_DEBUG_OPEN_CLOSE,
4614
		     "%s: %d threads %d, nodes %d (ref %d), refs %d, active transactions %d\n",
4615
		     __func__, proc->pid, threads, nodes, incoming_refs,
4616
		     outgoing_refs, active_transactions);
4617

4618
	binder_proc_dec_tmpref(proc);
4619 4620 4621 4622 4623 4624 4625 4626
}

static void binder_deferred_func(struct work_struct *work)
{
	struct binder_proc *proc;
	struct files_struct *files;

	int defer;
4627

4628 4629 4630 4631 4632 4633 4634 4635 4636 4637 4638 4639 4640 4641 4642 4643 4644 4645 4646 4647 4648 4649 4650 4651 4652 4653 4654 4655 4656 4657 4658 4659 4660 4661 4662 4663 4664 4665 4666 4667 4668
	do {
		mutex_lock(&binder_deferred_lock);
		if (!hlist_empty(&binder_deferred_list)) {
			proc = hlist_entry(binder_deferred_list.first,
					struct binder_proc, deferred_work_node);
			hlist_del_init(&proc->deferred_work_node);
			defer = proc->deferred_work;
			proc->deferred_work = 0;
		} else {
			proc = NULL;
			defer = 0;
		}
		mutex_unlock(&binder_deferred_lock);

		files = NULL;
		if (defer & BINDER_DEFERRED_PUT_FILES) {
			files = proc->files;
			if (files)
				proc->files = NULL;
		}

		if (defer & BINDER_DEFERRED_FLUSH)
			binder_deferred_flush(proc);

		if (defer & BINDER_DEFERRED_RELEASE)
			binder_deferred_release(proc); /* frees proc */

		if (files)
			put_files_struct(files);
	} while (proc);
}
static DECLARE_WORK(binder_deferred_work, binder_deferred_func);

static void
binder_defer_work(struct binder_proc *proc, enum binder_deferred_state defer)
{
	mutex_lock(&binder_deferred_lock);
	proc->deferred_work |= defer;
	if (hlist_unhashed(&proc->deferred_work_node)) {
		hlist_add_head(&proc->deferred_work_node,
				&binder_deferred_list);
4669
		schedule_work(&binder_deferred_work);
4670 4671 4672 4673
	}
	mutex_unlock(&binder_deferred_lock);
}

4674 4675 4676 4677
static void print_binder_transaction_ilocked(struct seq_file *m,
					     struct binder_proc *proc,
					     const char *prefix,
					     struct binder_transaction *t)
4678
{
4679 4680 4681 4682
	struct binder_proc *to_proc;
	struct binder_buffer *buffer = t->buffer;

	WARN_ON(!spin_is_locked(&proc->inner_lock));
4683
	spin_lock(&t->lock);
4684
	to_proc = t->to_proc;
4685 4686 4687 4688 4689
	seq_printf(m,
		   "%s %d: %p from %d:%d to %d:%d code %x flags %x pri %ld r%d",
		   prefix, t->debug_id, t,
		   t->from ? t->from->proc->pid : 0,
		   t->from ? t->from->pid : 0,
4690
		   to_proc ? to_proc->pid : 0,
4691 4692
		   t->to_thread ? t->to_thread->pid : 0,
		   t->code, t->flags, t->priority, t->need_reply);
4693 4694
	spin_unlock(&t->lock);

4695 4696 4697 4698 4699 4700 4701 4702 4703 4704
	if (proc != to_proc) {
		/*
		 * Can only safely deref buffer if we are holding the
		 * correct proc inner lock for this node
		 */
		seq_puts(m, "\n");
		return;
	}

	if (buffer == NULL) {
4705 4706
		seq_puts(m, " buffer free\n");
		return;
4707
	}
4708 4709
	if (buffer->target_node)
		seq_printf(m, " node %d", buffer->target_node->debug_id);
4710
	seq_printf(m, " size %zd:%zd data %p\n",
4711 4712
		   buffer->data_size, buffer->offsets_size,
		   buffer->data);
4713 4714
}

4715 4716 4717 4718 4719
static void print_binder_work_ilocked(struct seq_file *m,
				     struct binder_proc *proc,
				     const char *prefix,
				     const char *transaction_prefix,
				     struct binder_work *w)
4720 4721 4722 4723 4724 4725 4726
{
	struct binder_node *node;
	struct binder_transaction *t;

	switch (w->type) {
	case BINDER_WORK_TRANSACTION:
		t = container_of(w, struct binder_transaction, work);
4727 4728
		print_binder_transaction_ilocked(
				m, proc, transaction_prefix, t);
4729
		break;
4730 4731 4732 4733 4734 4735 4736
	case BINDER_WORK_RETURN_ERROR: {
		struct binder_error *e = container_of(
				w, struct binder_error, work);

		seq_printf(m, "%stransaction error: %u\n",
			   prefix, e->cmd);
	} break;
4737
	case BINDER_WORK_TRANSACTION_COMPLETE:
4738
		seq_printf(m, "%stransaction complete\n", prefix);
4739 4740 4741
		break;
	case BINDER_WORK_NODE:
		node = container_of(w, struct binder_node, work);
4742 4743 4744
		seq_printf(m, "%snode work %d: u%016llx c%016llx\n",
			   prefix, node->debug_id,
			   (u64)node->ptr, (u64)node->cookie);
4745 4746
		break;
	case BINDER_WORK_DEAD_BINDER:
4747
		seq_printf(m, "%shas dead binder\n", prefix);
4748 4749
		break;
	case BINDER_WORK_DEAD_BINDER_AND_CLEAR:
4750
		seq_printf(m, "%shas cleared dead binder\n", prefix);
4751 4752
		break;
	case BINDER_WORK_CLEAR_DEATH_NOTIFICATION:
4753
		seq_printf(m, "%shas cleared death notification\n", prefix);
4754 4755
		break;
	default:
4756
		seq_printf(m, "%sunknown work: type %d\n", prefix, w->type);
4757 4758 4759 4760
		break;
	}
}

4761 4762 4763
static void print_binder_thread_ilocked(struct seq_file *m,
					struct binder_thread *thread,
					int print_always)
4764 4765 4766
{
	struct binder_transaction *t;
	struct binder_work *w;
4767 4768
	size_t start_pos = m->count;
	size_t header_pos;
4769

4770
	WARN_ON(!spin_is_locked(&thread->proc->inner_lock));
4771
	seq_printf(m, "  thread %d: l %02x need_return %d tr %d\n",
4772
			thread->pid, thread->looper,
4773 4774
			thread->looper_need_return,
			atomic_read(&thread->tmp_ref));
4775
	header_pos = m->count;
4776 4777 4778
	t = thread->transaction_stack;
	while (t) {
		if (t->from == thread) {
4779 4780
			print_binder_transaction_ilocked(m, thread->proc,
					"    outgoing transaction", t);
4781 4782
			t = t->from_parent;
		} else if (t->to_thread == thread) {
4783
			print_binder_transaction_ilocked(m, thread->proc,
4784
						 "    incoming transaction", t);
4785 4786
			t = t->to_parent;
		} else {
4787 4788
			print_binder_transaction_ilocked(m, thread->proc,
					"    bad transaction", t);
4789 4790 4791 4792
			t = NULL;
		}
	}
	list_for_each_entry(w, &thread->todo, entry) {
4793
		print_binder_work_ilocked(m, thread->proc, "    ",
4794
					  "    pending transaction", w);
4795
	}
4796 4797
	if (!print_always && m->count == header_pos)
		m->count = start_pos;
4798 4799
}

4800 4801
static void print_binder_node_nilocked(struct seq_file *m,
				       struct binder_node *node)
4802 4803 4804 4805 4806
{
	struct binder_ref *ref;
	struct binder_work *w;
	int count;

4807
	WARN_ON(!spin_is_locked(&node->lock));
4808 4809
	if (node->proc)
		WARN_ON(!spin_is_locked(&node->proc->inner_lock));
4810

4811
	count = 0;
4812
	hlist_for_each_entry(ref, &node->refs, node_entry)
4813 4814
		count++;

4815
	seq_printf(m, "  node %d: u%016llx c%016llx hs %d hw %d ls %d lw %d is %d iw %d tr %d",
4816
		   node->debug_id, (u64)node->ptr, (u64)node->cookie,
4817 4818
		   node->has_strong_ref, node->has_weak_ref,
		   node->local_strong_refs, node->local_weak_refs,
4819
		   node->internal_strong_refs, count, node->tmp_refs);
4820
	if (count) {
4821
		seq_puts(m, " proc");
4822
		hlist_for_each_entry(ref, &node->refs, node_entry)
4823
			seq_printf(m, " %d", ref->proc->pid);
4824
	}
4825
	seq_puts(m, "\n");
4826 4827
	if (node->proc) {
		list_for_each_entry(w, &node->async_todo, entry)
4828
			print_binder_work_ilocked(m, node->proc, "    ",
4829 4830
					  "    pending async transaction", w);
	}
4831 4832
}

4833 4834
static void print_binder_ref_olocked(struct seq_file *m,
				     struct binder_ref *ref)
4835
{
4836
	WARN_ON(!spin_is_locked(&ref->proc->outer_lock));
4837
	binder_node_lock(ref->node);
4838 4839 4840 4841 4842
	seq_printf(m, "  ref %d: desc %d %snode %d s %d w %d d %pK\n",
		   ref->data.debug_id, ref->data.desc,
		   ref->node->proc ? "" : "dead ",
		   ref->node->debug_id, ref->data.strong,
		   ref->data.weak, ref->death);
4843
	binder_node_unlock(ref->node);
4844 4845
}

4846 4847
static void print_binder_proc(struct seq_file *m,
			      struct binder_proc *proc, int print_all)
4848 4849 4850
{
	struct binder_work *w;
	struct rb_node *n;
4851 4852
	size_t start_pos = m->count;
	size_t header_pos;
4853
	struct binder_node *last_node = NULL;
4854 4855

	seq_printf(m, "proc %d\n", proc->pid);
4856
	seq_printf(m, "context %s\n", proc->context->name);
4857 4858
	header_pos = m->count;

4859
	binder_inner_proc_lock(proc);
4860
	for (n = rb_first(&proc->threads); n != NULL; n = rb_next(n))
4861
		print_binder_thread_ilocked(m, rb_entry(n, struct binder_thread,
4862
						rb_node), print_all);
4863

4864
	for (n = rb_first(&proc->nodes); n != NULL; n = rb_next(n)) {
4865 4866
		struct binder_node *node = rb_entry(n, struct binder_node,
						    rb_node);
4867 4868 4869 4870 4871 4872 4873 4874 4875 4876 4877 4878 4879 4880 4881
		/*
		 * take a temporary reference on the node so it
		 * survives and isn't removed from the tree
		 * while we print it.
		 */
		binder_inc_node_tmpref_ilocked(node);
		/* Need to drop inner lock to take node lock */
		binder_inner_proc_unlock(proc);
		if (last_node)
			binder_put_node(last_node);
		binder_node_inner_lock(node);
		print_binder_node_nilocked(m, node);
		binder_node_inner_unlock(node);
		last_node = node;
		binder_inner_proc_lock(proc);
4882
	}
4883 4884 4885 4886
	binder_inner_proc_unlock(proc);
	if (last_node)
		binder_put_node(last_node);

4887
	if (print_all) {
4888
		binder_proc_lock(proc);
4889
		for (n = rb_first(&proc->refs_by_desc);
4890
		     n != NULL;
4891
		     n = rb_next(n))
4892 4893 4894 4895
			print_binder_ref_olocked(m, rb_entry(n,
							    struct binder_ref,
							    rb_node_desc));
		binder_proc_unlock(proc);
4896
	}
4897
	binder_alloc_print_allocated(m, &proc->alloc);
4898
	binder_inner_proc_lock(proc);
4899
	list_for_each_entry(w, &proc->todo, entry)
4900 4901
		print_binder_work_ilocked(m, proc, "  ",
					  "  pending transaction", w);
4902
	list_for_each_entry(w, &proc->delivered_death, entry) {
4903
		seq_puts(m, "  has delivered dead binder\n");
4904 4905
		break;
	}
4906
	binder_inner_proc_unlock(proc);
4907 4908
	if (!print_all && m->count == header_pos)
		m->count = start_pos;
4909 4910
}

4911
static const char * const binder_return_strings[] = {
4912 4913 4914 4915 4916 4917 4918 4919 4920 4921 4922 4923 4924 4925 4926 4927 4928 4929 4930 4931
	"BR_ERROR",
	"BR_OK",
	"BR_TRANSACTION",
	"BR_REPLY",
	"BR_ACQUIRE_RESULT",
	"BR_DEAD_REPLY",
	"BR_TRANSACTION_COMPLETE",
	"BR_INCREFS",
	"BR_ACQUIRE",
	"BR_RELEASE",
	"BR_DECREFS",
	"BR_ATTEMPT_ACQUIRE",
	"BR_NOOP",
	"BR_SPAWN_LOOPER",
	"BR_FINISHED",
	"BR_DEAD_BINDER",
	"BR_CLEAR_DEATH_NOTIFICATION_DONE",
	"BR_FAILED_REPLY"
};

4932
static const char * const binder_command_strings[] = {
4933 4934 4935 4936 4937 4938 4939 4940 4941 4942 4943 4944 4945 4946 4947 4948
	"BC_TRANSACTION",
	"BC_REPLY",
	"BC_ACQUIRE_RESULT",
	"BC_FREE_BUFFER",
	"BC_INCREFS",
	"BC_ACQUIRE",
	"BC_RELEASE",
	"BC_DECREFS",
	"BC_INCREFS_DONE",
	"BC_ACQUIRE_DONE",
	"BC_ATTEMPT_ACQUIRE",
	"BC_REGISTER_LOOPER",
	"BC_ENTER_LOOPER",
	"BC_EXIT_LOOPER",
	"BC_REQUEST_DEATH_NOTIFICATION",
	"BC_CLEAR_DEATH_NOTIFICATION",
4949 4950 4951
	"BC_DEAD_BINDER_DONE",
	"BC_TRANSACTION_SG",
	"BC_REPLY_SG",
4952 4953
};

4954
static const char * const binder_objstat_strings[] = {
4955 4956 4957 4958 4959 4960 4961 4962 4963
	"proc",
	"thread",
	"node",
	"ref",
	"death",
	"transaction",
	"transaction_complete"
};

4964 4965
static void print_binder_stats(struct seq_file *m, const char *prefix,
			       struct binder_stats *stats)
4966 4967 4968 4969
{
	int i;

	BUILD_BUG_ON(ARRAY_SIZE(stats->bc) !=
4970
		     ARRAY_SIZE(binder_command_strings));
4971
	for (i = 0; i < ARRAY_SIZE(stats->bc); i++) {
4972 4973 4974
		int temp = atomic_read(&stats->bc[i]);

		if (temp)
4975
			seq_printf(m, "%s%s: %d\n", prefix,
4976
				   binder_command_strings[i], temp);
4977 4978 4979
	}

	BUILD_BUG_ON(ARRAY_SIZE(stats->br) !=
4980
		     ARRAY_SIZE(binder_return_strings));
4981
	for (i = 0; i < ARRAY_SIZE(stats->br); i++) {
4982 4983 4984
		int temp = atomic_read(&stats->br[i]);

		if (temp)
4985
			seq_printf(m, "%s%s: %d\n", prefix,
4986
				   binder_return_strings[i], temp);
4987 4988 4989
	}

	BUILD_BUG_ON(ARRAY_SIZE(stats->obj_created) !=
4990
		     ARRAY_SIZE(binder_objstat_strings));
4991
	BUILD_BUG_ON(ARRAY_SIZE(stats->obj_created) !=
4992
		     ARRAY_SIZE(stats->obj_deleted));
4993
	for (i = 0; i < ARRAY_SIZE(stats->obj_created); i++) {
4994 4995 4996 4997 4998 4999
		int created = atomic_read(&stats->obj_created[i]);
		int deleted = atomic_read(&stats->obj_deleted[i]);

		if (created || deleted)
			seq_printf(m, "%s%s: active %d total %d\n",
				prefix,
5000
				binder_objstat_strings[i],
5001 5002
				created - deleted,
				created);
5003 5004 5005
	}
}

5006 5007
static void print_binder_proc_stats(struct seq_file *m,
				    struct binder_proc *proc)
5008 5009 5010 5011
{
	struct binder_work *w;
	struct rb_node *n;
	int count, strong, weak;
5012 5013
	size_t free_async_space =
		binder_alloc_get_free_async_space(&proc->alloc);
5014

5015
	seq_printf(m, "proc %d\n", proc->pid);
5016
	seq_printf(m, "context %s\n", proc->context->name);
5017
	count = 0;
5018
	binder_inner_proc_lock(proc);
5019 5020
	for (n = rb_first(&proc->threads); n != NULL; n = rb_next(n))
		count++;
5021 5022
	seq_printf(m, "  threads: %d\n", count);
	seq_printf(m, "  requested threads: %d+%d/%d\n"
5023 5024 5025
			"  ready threads %d\n"
			"  free async space %zd\n", proc->requested_threads,
			proc->requested_threads_started, proc->max_threads,
5026
			proc->ready_threads,
5027
			free_async_space);
5028 5029 5030
	count = 0;
	for (n = rb_first(&proc->nodes); n != NULL; n = rb_next(n))
		count++;
5031
	binder_inner_proc_unlock(proc);
5032
	seq_printf(m, "  nodes: %d\n", count);
5033 5034 5035
	count = 0;
	strong = 0;
	weak = 0;
5036
	binder_proc_lock(proc);
5037 5038 5039 5040
	for (n = rb_first(&proc->refs_by_desc); n != NULL; n = rb_next(n)) {
		struct binder_ref *ref = rb_entry(n, struct binder_ref,
						  rb_node_desc);
		count++;
5041 5042
		strong += ref->data.strong;
		weak += ref->data.weak;
5043
	}
5044
	binder_proc_unlock(proc);
5045
	seq_printf(m, "  refs: %d s %d w %d\n", count, strong, weak);
5046

5047
	count = binder_alloc_get_allocated_count(&proc->alloc);
5048
	seq_printf(m, "  buffers: %d\n", count);
5049 5050

	count = 0;
5051
	binder_inner_proc_lock(proc);
5052
	list_for_each_entry(w, &proc->todo, entry) {
5053
		if (w->type == BINDER_WORK_TRANSACTION)
5054 5055
			count++;
	}
5056
	binder_inner_proc_unlock(proc);
5057
	seq_printf(m, "  pending transactions: %d\n", count);
5058

5059
	print_binder_stats(m, "  ", &proc->stats);
5060 5061 5062
}


5063
static int binder_state_show(struct seq_file *m, void *unused)
5064 5065 5066
{
	struct binder_proc *proc;
	struct binder_node *node;
5067
	struct binder_node *last_node = NULL;
5068

5069
	seq_puts(m, "binder state:\n");
5070

5071
	spin_lock(&binder_dead_nodes_lock);
5072
	if (!hlist_empty(&binder_dead_nodes))
5073
		seq_puts(m, "dead nodes:\n");
5074 5075 5076 5077 5078 5079 5080 5081 5082 5083 5084
	hlist_for_each_entry(node, &binder_dead_nodes, dead_node) {
		/*
		 * take a temporary reference on the node so it
		 * survives and isn't removed from the list
		 * while we print it.
		 */
		node->tmp_refs++;
		spin_unlock(&binder_dead_nodes_lock);
		if (last_node)
			binder_put_node(last_node);
		binder_node_lock(node);
5085
		print_binder_node_nilocked(m, node);
5086 5087 5088 5089
		binder_node_unlock(node);
		last_node = node;
		spin_lock(&binder_dead_nodes_lock);
	}
5090
	spin_unlock(&binder_dead_nodes_lock);
5091 5092
	if (last_node)
		binder_put_node(last_node);
5093

5094
	mutex_lock(&binder_procs_lock);
5095
	hlist_for_each_entry(proc, &binder_procs, proc_node)
5096
		print_binder_proc(m, proc, 1);
5097
	mutex_unlock(&binder_procs_lock);
T
Todd Kjos 已提交
5098

5099
	return 0;
5100 5101
}

5102
static int binder_stats_show(struct seq_file *m, void *unused)
5103 5104 5105
{
	struct binder_proc *proc;

5106
	seq_puts(m, "binder stats:\n");
5107

5108
	print_binder_stats(m, "", &binder_stats);
5109

5110
	mutex_lock(&binder_procs_lock);
5111
	hlist_for_each_entry(proc, &binder_procs, proc_node)
5112
		print_binder_proc_stats(m, proc);
5113
	mutex_unlock(&binder_procs_lock);
T
Todd Kjos 已提交
5114

5115
	return 0;
5116 5117
}

5118
static int binder_transactions_show(struct seq_file *m, void *unused)
5119 5120 5121
{
	struct binder_proc *proc;

5122
	seq_puts(m, "binder transactions:\n");
5123
	mutex_lock(&binder_procs_lock);
5124
	hlist_for_each_entry(proc, &binder_procs, proc_node)
5125
		print_binder_proc(m, proc, 0);
5126
	mutex_unlock(&binder_procs_lock);
T
Todd Kjos 已提交
5127

5128
	return 0;
5129 5130
}

5131
static int binder_proc_show(struct seq_file *m, void *unused)
5132
{
5133
	struct binder_proc *itr;
5134
	int pid = (unsigned long)m->private;
5135

5136
	mutex_lock(&binder_procs_lock);
5137
	hlist_for_each_entry(itr, &binder_procs, proc_node) {
5138 5139 5140
		if (itr->pid == pid) {
			seq_puts(m, "binder proc state:\n");
			print_binder_proc(m, itr, 1);
5141 5142
		}
	}
5143 5144
	mutex_unlock(&binder_procs_lock);

5145
	return 0;
5146 5147
}

5148
static void print_binder_transaction_log_entry(struct seq_file *m,
5149 5150
					struct binder_transaction_log_entry *e)
{
5151 5152 5153 5154 5155 5156
	int debug_id = READ_ONCE(e->debug_id_done);
	/*
	 * read barrier to guarantee debug_id_done read before
	 * we print the log values
	 */
	smp_rmb();
5157
	seq_printf(m,
5158
		   "%d: %s from %d:%d to %d:%d context %s node %d handle %d size %d:%d ret %d/%d l=%d",
5159 5160
		   e->debug_id, (e->call_type == 2) ? "reply" :
		   ((e->call_type == 1) ? "async" : "call "), e->from_proc,
5161
		   e->from_thread, e->to_proc, e->to_thread, e->context_name,
5162 5163 5164
		   e->to_node, e->target_handle, e->data_size, e->offsets_size,
		   e->return_error, e->return_error_param,
		   e->return_error_line);
5165 5166 5167 5168 5169 5170 5171
	/*
	 * read-barrier to guarantee read of debug_id_done after
	 * done printing the fields of the entry
	 */
	smp_rmb();
	seq_printf(m, debug_id && debug_id == READ_ONCE(e->debug_id_done) ?
			"\n" : " (incomplete)\n");
5172 5173
}

5174
static int binder_transaction_log_show(struct seq_file *m, void *unused)
5175
{
5176
	struct binder_transaction_log *log = m->private;
5177 5178 5179
	unsigned int log_cur = atomic_read(&log->cur);
	unsigned int count;
	unsigned int cur;
5180 5181
	int i;

5182 5183 5184 5185 5186 5187 5188 5189 5190
	count = log_cur + 1;
	cur = count < ARRAY_SIZE(log->entry) && !log->full ?
		0 : count % ARRAY_SIZE(log->entry);
	if (count > ARRAY_SIZE(log->entry) || log->full)
		count = ARRAY_SIZE(log->entry);
	for (i = 0; i < count; i++) {
		unsigned int index = cur++ % ARRAY_SIZE(log->entry);

		print_binder_transaction_log_entry(m, &log->entry[index]);
5191
	}
5192
	return 0;
5193 5194 5195 5196 5197 5198
}

static const struct file_operations binder_fops = {
	.owner = THIS_MODULE,
	.poll = binder_poll,
	.unlocked_ioctl = binder_ioctl,
5199
	.compat_ioctl = binder_ioctl,
5200 5201 5202 5203 5204 5205
	.mmap = binder_mmap,
	.open = binder_open,
	.flush = binder_flush,
	.release = binder_release,
};

5206 5207 5208 5209 5210
BINDER_DEBUG_ENTRY(state);
BINDER_DEBUG_ENTRY(stats);
BINDER_DEBUG_ENTRY(transactions);
BINDER_DEBUG_ENTRY(transaction_log);

5211 5212 5213 5214 5215 5216 5217 5218 5219 5220 5221 5222 5223 5224 5225
static int __init init_binder_device(const char *name)
{
	int ret;
	struct binder_device *binder_device;

	binder_device = kzalloc(sizeof(*binder_device), GFP_KERNEL);
	if (!binder_device)
		return -ENOMEM;

	binder_device->miscdev.fops = &binder_fops;
	binder_device->miscdev.minor = MISC_DYNAMIC_MINOR;
	binder_device->miscdev.name = name;

	binder_device->context.binder_context_mgr_uid = INVALID_UID;
	binder_device->context.name = name;
5226
	mutex_init(&binder_device->context.context_mgr_node_lock);
5227 5228 5229 5230 5231 5232 5233 5234 5235 5236 5237 5238

	ret = misc_register(&binder_device->miscdev);
	if (ret < 0) {
		kfree(binder_device);
		return ret;
	}

	hlist_add_head(&binder_device->hlist, &binder_devices);

	return ret;
}

5239 5240 5241
static int __init binder_init(void)
{
	int ret;
C
Christian Brauner 已提交
5242
	char *device_name, *device_names, *device_tmp;
5243 5244
	struct binder_device *device;
	struct hlist_node *tmp;
5245

5246 5247 5248
	atomic_set(&binder_transaction_log.cur, ~0U);
	atomic_set(&binder_transaction_log_failed.cur, ~0U);

5249 5250 5251 5252
	binder_debugfs_dir_entry_root = debugfs_create_dir("binder", NULL);
	if (binder_debugfs_dir_entry_root)
		binder_debugfs_dir_entry_proc = debugfs_create_dir("proc",
						 binder_debugfs_dir_entry_root);
5253

5254 5255 5256 5257 5258 5259 5260 5261 5262 5263 5264 5265 5266 5267 5268 5269 5270 5271 5272 5273 5274 5275 5276 5277 5278 5279
	if (binder_debugfs_dir_entry_root) {
		debugfs_create_file("state",
				    S_IRUGO,
				    binder_debugfs_dir_entry_root,
				    NULL,
				    &binder_state_fops);
		debugfs_create_file("stats",
				    S_IRUGO,
				    binder_debugfs_dir_entry_root,
				    NULL,
				    &binder_stats_fops);
		debugfs_create_file("transactions",
				    S_IRUGO,
				    binder_debugfs_dir_entry_root,
				    NULL,
				    &binder_transactions_fops);
		debugfs_create_file("transaction_log",
				    S_IRUGO,
				    binder_debugfs_dir_entry_root,
				    &binder_transaction_log,
				    &binder_transaction_log_fops);
		debugfs_create_file("failed_transaction_log",
				    S_IRUGO,
				    binder_debugfs_dir_entry_root,
				    &binder_transaction_log_failed,
				    &binder_transaction_log_fops);
5280
	}
5281 5282 5283 5284 5285 5286 5287 5288 5289 5290 5291 5292

	/*
	 * Copy the module_parameter string, because we don't want to
	 * tokenize it in-place.
	 */
	device_names = kzalloc(strlen(binder_devices_param) + 1, GFP_KERNEL);
	if (!device_names) {
		ret = -ENOMEM;
		goto err_alloc_device_names_failed;
	}
	strcpy(device_names, binder_devices_param);

C
Christian Brauner 已提交
5293 5294
	device_tmp = device_names;
	while ((device_name = strsep(&device_tmp, ","))) {
5295 5296 5297 5298 5299 5300 5301 5302 5303 5304 5305 5306 5307
		ret = init_binder_device(device_name);
		if (ret)
			goto err_init_binder_device_failed;
	}

	return ret;

err_init_binder_device_failed:
	hlist_for_each_entry_safe(device, tmp, &binder_devices, hlist) {
		misc_deregister(&device->miscdev);
		hlist_del(&device->hlist);
		kfree(device);
	}
C
Christian Brauner 已提交
5308 5309 5310

	kfree(device_names);

5311 5312 5313
err_alloc_device_names_failed:
	debugfs_remove_recursive(binder_debugfs_dir_entry_root);

5314 5315 5316 5317 5318
	return ret;
}

device_initcall(binder_init);

5319 5320 5321
#define CREATE_TRACE_POINTS
#include "binder_trace.h"

5322
MODULE_LICENSE("GPL v2");