binder_alloc.c 35.1 KB
Newer Older
1
// SPDX-License-Identifier: GPL-2.0-only
2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
/* binder_alloc.c
 *
 * Android IPC Subsystem
 *
 * Copyright (C) 2007-2017 Google, Inc.
 */

#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt

#include <linux/list.h>
#include <linux/sched/mm.h>
#include <linux/module.h>
#include <linux/rtmutex.h>
#include <linux/rbtree.h>
#include <linux/seq_file.h>
#include <linux/vmalloc.h>
#include <linux/slab.h>
#include <linux/sched.h>
20
#include <linux/list_lru.h>
21
#include <linux/ratelimit.h>
22
#include <asm/cacheflush.h>
23 24
#include <linux/uaccess.h>
#include <linux/highmem.h>
25
#include <linux/sizes.h>
26 27 28
#include "binder_alloc.h"
#include "binder_trace.h"

29 30
struct list_lru binder_alloc_lru;

31 32 33
static DEFINE_MUTEX(binder_alloc_mmap_lock);

enum {
34
	BINDER_DEBUG_USER_ERROR             = 1U << 0,
35 36 37 38
	BINDER_DEBUG_OPEN_CLOSE             = 1U << 1,
	BINDER_DEBUG_BUFFER_ALLOC           = 1U << 2,
	BINDER_DEBUG_BUFFER_ALLOC_ASYNC     = 1U << 3,
};
39
static uint32_t binder_alloc_debug_mask = BINDER_DEBUG_USER_ERROR;
40 41 42 43 44 45 46

module_param_named(debug_mask, binder_alloc_debug_mask,
		   uint, 0644);

#define binder_alloc_debug(mask, x...) \
	do { \
		if (binder_alloc_debug_mask & mask) \
47
			pr_info_ratelimited(x); \
48 49
	} while (0)

50 51 52 53 54 55 56 57 58 59
static struct binder_buffer *binder_buffer_next(struct binder_buffer *buffer)
{
	return list_entry(buffer->entry.next, struct binder_buffer, entry);
}

static struct binder_buffer *binder_buffer_prev(struct binder_buffer *buffer)
{
	return list_entry(buffer->entry.prev, struct binder_buffer, entry);
}

60 61 62 63
static size_t binder_alloc_buffer_size(struct binder_alloc *alloc,
				       struct binder_buffer *buffer)
{
	if (list_is_last(&buffer->entry, &alloc->buffers))
64 65
		return alloc->buffer + alloc->buffer_size - buffer->user_data;
	return binder_buffer_next(buffer)->user_data - buffer->user_data;
66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114
}

static void binder_insert_free_buffer(struct binder_alloc *alloc,
				      struct binder_buffer *new_buffer)
{
	struct rb_node **p = &alloc->free_buffers.rb_node;
	struct rb_node *parent = NULL;
	struct binder_buffer *buffer;
	size_t buffer_size;
	size_t new_buffer_size;

	BUG_ON(!new_buffer->free);

	new_buffer_size = binder_alloc_buffer_size(alloc, new_buffer);

	binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC,
		     "%d: add free buffer, size %zd, at %pK\n",
		      alloc->pid, new_buffer_size, new_buffer);

	while (*p) {
		parent = *p;
		buffer = rb_entry(parent, struct binder_buffer, rb_node);
		BUG_ON(!buffer->free);

		buffer_size = binder_alloc_buffer_size(alloc, buffer);

		if (new_buffer_size < buffer_size)
			p = &parent->rb_left;
		else
			p = &parent->rb_right;
	}
	rb_link_node(&new_buffer->rb_node, parent, p);
	rb_insert_color(&new_buffer->rb_node, &alloc->free_buffers);
}

static void binder_insert_allocated_buffer_locked(
		struct binder_alloc *alloc, struct binder_buffer *new_buffer)
{
	struct rb_node **p = &alloc->allocated_buffers.rb_node;
	struct rb_node *parent = NULL;
	struct binder_buffer *buffer;

	BUG_ON(new_buffer->free);

	while (*p) {
		parent = *p;
		buffer = rb_entry(parent, struct binder_buffer, rb_node);
		BUG_ON(buffer->free);

115
		if (new_buffer->user_data < buffer->user_data)
116
			p = &parent->rb_left;
117
		else if (new_buffer->user_data > buffer->user_data)
118 119 120 121 122 123 124 125
			p = &parent->rb_right;
		else
			BUG();
	}
	rb_link_node(&new_buffer->rb_node, parent, p);
	rb_insert_color(&new_buffer->rb_node, &alloc->allocated_buffers);
}

126
static struct binder_buffer *binder_alloc_prepare_to_free_locked(
127 128 129 130 131
		struct binder_alloc *alloc,
		uintptr_t user_ptr)
{
	struct rb_node *n = alloc->allocated_buffers.rb_node;
	struct binder_buffer *buffer;
132
	void __user *uptr;
133

134
	uptr = (void __user *)user_ptr;
135 136 137 138 139

	while (n) {
		buffer = rb_entry(n, struct binder_buffer, rb_node);
		BUG_ON(buffer->free);

140
		if (uptr < buffer->user_data)
141
			n = n->rb_left;
142
		else if (uptr > buffer->user_data)
143
			n = n->rb_right;
144 145 146
		else {
			/*
			 * Guard against user threads attempting to
147 148
			 * free the buffer when in use by kernel or
			 * after it's already been freed.
149
			 */
150 151 152
			if (!buffer->allow_user_free)
				return ERR_PTR(-EPERM);
			buffer->allow_user_free = 0;
153
			return buffer;
154
		}
155 156 157 158 159
	}
	return NULL;
}

/**
160
 * binder_alloc_prepare_to_free() - get buffer given user ptr
161 162 163 164 165 166 167 168 169
 * @alloc:	binder_alloc for this proc
 * @user_ptr:	User pointer to buffer data
 *
 * Validate userspace pointer to buffer data and return buffer corresponding to
 * that user pointer. Search the rb tree for buffer that matches user data
 * pointer.
 *
 * Return:	Pointer to buffer or NULL
 */
170 171
struct binder_buffer *binder_alloc_prepare_to_free(struct binder_alloc *alloc,
						   uintptr_t user_ptr)
172 173 174 175
{
	struct binder_buffer *buffer;

	mutex_lock(&alloc->mutex);
176
	buffer = binder_alloc_prepare_to_free_locked(alloc, user_ptr);
177 178 179 180 181
	mutex_unlock(&alloc->mutex);
	return buffer;
}

static int binder_update_page_range(struct binder_alloc *alloc, int allocate,
182
				    void __user *start, void __user *end)
183
{
184
	void __user *page_addr;
185
	unsigned long user_page_addr;
186
	struct binder_lru_page *page;
187
	struct vm_area_struct *vma = NULL;
188 189
	struct mm_struct *mm = NULL;
	bool need_mm = false;
190 191 192 193 194 195 196 197 198 199

	binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC,
		     "%d: %s pages %pK-%pK\n", alloc->pid,
		     allocate ? "allocate" : "free", start, end);

	if (end <= start)
		return 0;

	trace_binder_update_page_range(alloc, allocate, start, end);

200 201 202 203 204 205 206 207 208 209 210
	if (allocate == 0)
		goto free_range;

	for (page_addr = start; page_addr < end; page_addr += PAGE_SIZE) {
		page = &alloc->pages[(page_addr - alloc->buffer) / PAGE_SIZE];
		if (!page->page_ptr) {
			need_mm = true;
			break;
		}
	}

211
	if (need_mm && mmget_not_zero(alloc->vma_vm_mm))
212
		mm = alloc->vma_vm_mm;
213 214

	if (mm) {
215
		mmap_read_lock(mm);
216
		vma = vma_lookup(mm, alloc->vma_addr);
217 218
	}

219
	if (!vma && need_mm) {
220 221 222
		binder_alloc_debug(BINDER_DEBUG_USER_ERROR,
				   "%d: binder_alloc_buf failed to map pages in userspace, no vma\n",
				   alloc->pid);
223 224 225 226 227
		goto err_no_vma;
	}

	for (page_addr = start; page_addr < end; page_addr += PAGE_SIZE) {
		int ret;
228
		bool on_lru;
229
		size_t index;
230

231 232
		index = (page_addr - alloc->buffer) / PAGE_SIZE;
		page = &alloc->pages[index];
233

234
		if (page->page_ptr) {
235 236
			trace_binder_alloc_lru_start(alloc, index);

237 238
			on_lru = list_lru_del(&binder_alloc_lru, &page->lru);
			WARN_ON(!on_lru);
239 240

			trace_binder_alloc_lru_end(alloc, index);
241 242 243 244 245 246
			continue;
		}

		if (WARN_ON(!vma))
			goto err_page_ptr_cleared;

247
		trace_binder_alloc_page_start(alloc, index);
248 249 250 251
		page->page_ptr = alloc_page(GFP_KERNEL |
					    __GFP_HIGHMEM |
					    __GFP_ZERO);
		if (!page->page_ptr) {
252 253 254 255
			pr_err("%d: binder_alloc_buf failed for page at %pK\n",
				alloc->pid, page_addr);
			goto err_alloc_page_failed;
		}
256 257 258
		page->alloc = alloc;
		INIT_LIST_HEAD(&page->lru);

T
Todd Kjos 已提交
259
		user_page_addr = (uintptr_t)page_addr;
260
		ret = vm_insert_page(vma, user_page_addr, page[0].page_ptr);
261 262 263 264 265
		if (ret) {
			pr_err("%d: binder_alloc_buf failed to map page at %lx in userspace\n",
			       alloc->pid, user_page_addr);
			goto err_vm_insert_page_failed;
		}
266

267 268 269
		if (index + 1 > alloc->pages_high)
			alloc->pages_high = index + 1;

270
		trace_binder_alloc_page_end(alloc, index);
271 272
	}
	if (mm) {
273
		mmap_read_unlock(mm);
274 275 276 277 278
		mmput(mm);
	}
	return 0;

free_range:
279
	for (page_addr = end - PAGE_SIZE; 1; page_addr -= PAGE_SIZE) {
280
		bool ret;
281
		size_t index;
282

283 284 285 286
		index = (page_addr - alloc->buffer) / PAGE_SIZE;
		page = &alloc->pages[index];

		trace_binder_free_lru_start(alloc, index);
287 288 289

		ret = list_lru_add(&binder_alloc_lru, &page->lru);
		WARN_ON(!ret);
290 291

		trace_binder_free_lru_end(alloc, index);
292 293
		if (page_addr == start)
			break;
294 295
		continue;

296
err_vm_insert_page_failed:
297 298
		__free_page(page->page_ptr);
		page->page_ptr = NULL;
299
err_alloc_page_failed:
300
err_page_ptr_cleared:
301 302
		if (page_addr == start)
			break;
303 304 305
	}
err_no_vma:
	if (mm) {
306
		mmap_read_unlock(mm);
307 308
		mmput(mm);
	}
309
	return vma ? -ENOMEM : -ESRCH;
310 311
}

312 313 314 315

static inline void binder_alloc_set_vma(struct binder_alloc *alloc,
		struct vm_area_struct *vma)
{
316 317
	unsigned long vm_start = 0;

318 319 320 321 322
	/*
	 * Allow clearing the vma with holding just the read lock to allow
	 * munmapping downgrade of the write lock before freeing and closing the
	 * file using binder_alloc_vma_close().
	 */
323 324
	if (vma) {
		vm_start = vma->vm_start;
325 326 327
		mmap_assert_write_locked(alloc->vma_vm_mm);
	} else {
		mmap_assert_locked(alloc->vma_vm_mm);
328 329 330
	}

	alloc->vma_addr = vm_start;
331 332 333 334 335 336 337
}

static inline struct vm_area_struct *binder_alloc_get_vma(
		struct binder_alloc *alloc)
{
	struct vm_area_struct *vma = NULL;

338 339 340
	if (alloc->vma_addr)
		vma = vma_lookup(alloc->vma_vm_mm, alloc->vma_addr);

341 342 343
	return vma;
}

344
static bool debug_low_async_space_locked(struct binder_alloc *alloc, int pid)
345 346 347 348 349 350 351 352
{
	/*
	 * Find the amount and size of buffers allocated by the current caller;
	 * The idea is that once we cross the threshold, whoever is responsible
	 * for the low async space is likely to try to send another async txn,
	 * and at some point we'll catch them in the act. This is more efficient
	 * than keeping a map per pid.
	 */
353
	struct rb_node *n;
354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371
	struct binder_buffer *buffer;
	size_t total_alloc_size = 0;
	size_t num_buffers = 0;

	for (n = rb_first(&alloc->allocated_buffers); n != NULL;
		 n = rb_next(n)) {
		buffer = rb_entry(n, struct binder_buffer, rb_node);
		if (buffer->pid != pid)
			continue;
		if (!buffer->async_transaction)
			continue;
		total_alloc_size += binder_alloc_buffer_size(alloc, buffer)
			+ sizeof(struct binder_buffer);
		num_buffers++;
	}

	/*
	 * Warn if this pid has more than 50 transactions, or more than 50% of
372 373
	 * async space (which is 25% of total buffer size). Oneway spam is only
	 * detected when the threshold is exceeded.
374 375 376 377 378
	 */
	if (num_buffers > 50 || total_alloc_size > alloc->buffer_size / 4) {
		binder_alloc_debug(BINDER_DEBUG_USER_ERROR,
			     "%d: pid %d spamming oneway? %zd buffers allocated for a total size of %zd\n",
			      alloc->pid, pid, num_buffers, total_alloc_size);
379 380 381 382
		if (!alloc->oneway_spam_detected) {
			alloc->oneway_spam_detected = true;
			return true;
		}
383
	}
384
	return false;
385 386
}

387 388 389 390 391
static struct binder_buffer *binder_alloc_new_buf_locked(
				struct binder_alloc *alloc,
				size_t data_size,
				size_t offsets_size,
				size_t extra_buffers_size,
392 393
				int is_async,
				int pid)
394 395 396 397 398
{
	struct rb_node *n = alloc->free_buffers.rb_node;
	struct binder_buffer *buffer;
	size_t buffer_size;
	struct rb_node *best_fit = NULL;
399 400
	void __user *has_page_addr;
	void __user *end_page_addr;
401
	size_t size, data_offsets_size;
402
	int ret;
403

404
	mmap_read_lock(alloc->vma_vm_mm);
405
	if (!binder_alloc_get_vma(alloc)) {
406
		mmap_read_unlock(alloc->vma_vm_mm);
407 408 409
		binder_alloc_debug(BINDER_DEBUG_USER_ERROR,
				   "%d: binder_alloc_buf, no vma\n",
				   alloc->pid);
410
		return ERR_PTR(-ESRCH);
411
	}
412
	mmap_read_unlock(alloc->vma_vm_mm);
413 414 415 416 417 418 419 420

	data_offsets_size = ALIGN(data_size, sizeof(void *)) +
		ALIGN(offsets_size, sizeof(void *));

	if (data_offsets_size < data_size || data_offsets_size < offsets_size) {
		binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC,
				"%d: got transaction with invalid size %zd-%zd\n",
				alloc->pid, data_size, offsets_size);
421
		return ERR_PTR(-EINVAL);
422 423 424 425 426 427
	}
	size = data_offsets_size + ALIGN(extra_buffers_size, sizeof(void *));
	if (size < data_offsets_size || size < extra_buffers_size) {
		binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC,
				"%d: got transaction with invalid extra_buffers_size %zd\n",
				alloc->pid, extra_buffers_size);
428
		return ERR_PTR(-EINVAL);
429 430 431 432 433 434
	}
	if (is_async &&
	    alloc->free_async_space < size + sizeof(struct binder_buffer)) {
		binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC,
			     "%d: binder_alloc_buf size %zd failed, no async space left\n",
			      alloc->pid, size);
435
		return ERR_PTR(-ENOSPC);
436 437
	}

438 439 440
	/* Pad 0-size buffers so they get assigned unique addresses */
	size = max(size, sizeof(void *));

441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456
	while (n) {
		buffer = rb_entry(n, struct binder_buffer, rb_node);
		BUG_ON(!buffer->free);
		buffer_size = binder_alloc_buffer_size(alloc, buffer);

		if (size < buffer_size) {
			best_fit = n;
			n = n->rb_left;
		} else if (size > buffer_size)
			n = n->rb_right;
		else {
			best_fit = n;
			break;
		}
	}
	if (best_fit == NULL) {
457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481
		size_t allocated_buffers = 0;
		size_t largest_alloc_size = 0;
		size_t total_alloc_size = 0;
		size_t free_buffers = 0;
		size_t largest_free_size = 0;
		size_t total_free_size = 0;

		for (n = rb_first(&alloc->allocated_buffers); n != NULL;
		     n = rb_next(n)) {
			buffer = rb_entry(n, struct binder_buffer, rb_node);
			buffer_size = binder_alloc_buffer_size(alloc, buffer);
			allocated_buffers++;
			total_alloc_size += buffer_size;
			if (buffer_size > largest_alloc_size)
				largest_alloc_size = buffer_size;
		}
		for (n = rb_first(&alloc->free_buffers); n != NULL;
		     n = rb_next(n)) {
			buffer = rb_entry(n, struct binder_buffer, rb_node);
			buffer_size = binder_alloc_buffer_size(alloc, buffer);
			free_buffers++;
			total_free_size += buffer_size;
			if (buffer_size > largest_free_size)
				largest_free_size = buffer_size;
		}
482 483 484 485 486 487 488 489
		binder_alloc_debug(BINDER_DEBUG_USER_ERROR,
				   "%d: binder_alloc_buf size %zd failed, no address space\n",
				   alloc->pid, size);
		binder_alloc_debug(BINDER_DEBUG_USER_ERROR,
				   "allocated: %zd (num: %zd largest: %zd), free: %zd (num: %zd largest: %zd)\n",
				   total_alloc_size, allocated_buffers,
				   largest_alloc_size, total_free_size,
				   free_buffers, largest_free_size);
490
		return ERR_PTR(-ENOSPC);
491 492 493 494 495 496 497 498 499 500
	}
	if (n == NULL) {
		buffer = rb_entry(best_fit, struct binder_buffer, rb_node);
		buffer_size = binder_alloc_buffer_size(alloc, buffer);
	}

	binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC,
		     "%d: binder_alloc_buf size %zd got buffer %pK size %zd\n",
		      alloc->pid, size, buffer, buffer_size);

501 502
	has_page_addr = (void __user *)
		(((uintptr_t)buffer->user_data + buffer_size) & PAGE_MASK);
503
	WARN_ON(n && buffer_size != size);
504
	end_page_addr =
505
		(void __user *)PAGE_ALIGN((uintptr_t)buffer->user_data + size);
506 507
	if (end_page_addr > has_page_addr)
		end_page_addr = has_page_addr;
508 509
	ret = binder_update_page_range(alloc, 1, (void __user *)
		PAGE_ALIGN((uintptr_t)buffer->user_data), end_page_addr);
510 511
	if (ret)
		return ERR_PTR(ret);
512 513

	if (buffer_size != size) {
514
		struct binder_buffer *new_buffer;
515

516 517 518 519 520 521
		new_buffer = kzalloc(sizeof(*buffer), GFP_KERNEL);
		if (!new_buffer) {
			pr_err("%s: %d failed to alloc new buffer struct\n",
			       __func__, alloc->pid);
			goto err_alloc_buf_struct_failed;
		}
522
		new_buffer->user_data = (u8 __user *)buffer->user_data + size;
523 524 525 526
		list_add(&new_buffer->entry, &buffer->entry);
		new_buffer->free = 1;
		binder_insert_free_buffer(alloc, new_buffer);
	}
527 528 529

	rb_erase(best_fit, &alloc->free_buffers);
	buffer->free = 0;
530
	buffer->allow_user_free = 0;
531
	binder_insert_allocated_buffer_locked(alloc, buffer);
532 533 534 535 536 537 538
	binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC,
		     "%d: binder_alloc_buf size %zd got %pK\n",
		      alloc->pid, size, buffer);
	buffer->data_size = data_size;
	buffer->offsets_size = offsets_size;
	buffer->async_transaction = is_async;
	buffer->extra_buffers_size = extra_buffers_size;
539
	buffer->pid = pid;
540
	buffer->oneway_spam_suspect = false;
541 542 543 544 545
	if (is_async) {
		alloc->free_async_space -= size + sizeof(struct binder_buffer);
		binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC_ASYNC,
			     "%d: binder_alloc_buf size %zd async free %zd\n",
			      alloc->pid, size, alloc->free_async_space);
546 547 548 549 550 551
		if (alloc->free_async_space < alloc->buffer_size / 10) {
			/*
			 * Start detecting spammers once we have less than 20%
			 * of async space left (which is less than 10% of total
			 * buffer size).
			 */
552 553 554
			buffer->oneway_spam_suspect = debug_low_async_space_locked(alloc, pid);
		} else {
			alloc->oneway_spam_detected = false;
555
		}
556 557
	}
	return buffer;
558 559

err_alloc_buf_struct_failed:
560 561
	binder_update_page_range(alloc, 0, (void __user *)
				 PAGE_ALIGN((uintptr_t)buffer->user_data),
562
				 end_page_addr);
563
	return ERR_PTR(-ENOMEM);
564 565 566 567 568 569 570 571 572
}

/**
 * binder_alloc_new_buf() - Allocate a new binder buffer
 * @alloc:              binder_alloc for this proc
 * @data_size:          size of user data buffer
 * @offsets_size:       user specified buffer offset
 * @extra_buffers_size: size of extra space for meta-data (eg, security context)
 * @is_async:           buffer for async transaction
573
 * @pid:				pid to attribute allocation to (used for debugging)
574 575 576 577 578 579 580 581 582 583 584 585
 *
 * Allocate a new buffer given the requested sizes. Returns
 * the kernel version of the buffer pointer. The size allocated
 * is the sum of the three given sizes (each rounded up to
 * pointer-sized boundary)
 *
 * Return:	The allocated buffer or %NULL if error
 */
struct binder_buffer *binder_alloc_new_buf(struct binder_alloc *alloc,
					   size_t data_size,
					   size_t offsets_size,
					   size_t extra_buffers_size,
586 587
					   int is_async,
					   int pid)
588 589 590 591 592
{
	struct binder_buffer *buffer;

	mutex_lock(&alloc->mutex);
	buffer = binder_alloc_new_buf_locked(alloc, data_size, offsets_size,
593
					     extra_buffers_size, is_async, pid);
594 595 596 597
	mutex_unlock(&alloc->mutex);
	return buffer;
}

598
static void __user *buffer_start_page(struct binder_buffer *buffer)
599
{
600
	return (void __user *)((uintptr_t)buffer->user_data & PAGE_MASK);
601 602
}

603
static void __user *prev_buffer_end_page(struct binder_buffer *buffer)
604
{
605 606
	return (void __user *)
		(((uintptr_t)(buffer->user_data) - 1) & PAGE_MASK);
607 608 609 610 611 612
}

static void binder_delete_free_buffer(struct binder_alloc *alloc,
				      struct binder_buffer *buffer)
{
	struct binder_buffer *prev, *next = NULL;
613
	bool to_free = true;
614

615
	BUG_ON(alloc->buffers.next == &buffer->entry);
616
	prev = binder_buffer_prev(buffer);
617
	BUG_ON(!prev->free);
618 619
	if (prev_buffer_end_page(prev) == buffer_start_page(buffer)) {
		to_free = false;
620
		binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC,
621
				   "%d: merge free, buffer %pK share page with %pK\n",
622 623
				   alloc->pid, buffer->user_data,
				   prev->user_data);
624 625 626
	}

	if (!list_is_last(&buffer->entry, &alloc->buffers)) {
627
		next = binder_buffer_next(buffer);
628 629
		if (buffer_start_page(next) == buffer_start_page(buffer)) {
			to_free = false;
630
			binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC,
631 632
					   "%d: merge free, buffer %pK share page with %pK\n",
					   alloc->pid,
633 634
					   buffer->user_data,
					   next->user_data);
635 636
		}
	}
637

638
	if (PAGE_ALIGNED(buffer->user_data)) {
639 640
		binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC,
				   "%d: merge free, buffer start %pK is page aligned\n",
641
				   alloc->pid, buffer->user_data);
642 643 644 645
		to_free = false;
	}

	if (to_free) {
646
		binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC,
647
				   "%d: merge free, buffer %pK do not share page with %pK or %pK\n",
648 649 650
				   alloc->pid, buffer->user_data,
				   prev->user_data,
				   next ? next->user_data : NULL);
651
		binder_update_page_range(alloc, 0, buffer_start_page(buffer),
652
					 buffer_start_page(buffer) + PAGE_SIZE);
653
	}
654 655
	list_del(&buffer->entry);
	kfree(buffer);
656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675
}

static void binder_free_buf_locked(struct binder_alloc *alloc,
				   struct binder_buffer *buffer)
{
	size_t size, buffer_size;

	buffer_size = binder_alloc_buffer_size(alloc, buffer);

	size = ALIGN(buffer->data_size, sizeof(void *)) +
		ALIGN(buffer->offsets_size, sizeof(void *)) +
		ALIGN(buffer->extra_buffers_size, sizeof(void *));

	binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC,
		     "%d: binder_free_buf %pK size %zd buffer_size %zd\n",
		      alloc->pid, buffer, size, buffer_size);

	BUG_ON(buffer->free);
	BUG_ON(size > buffer_size);
	BUG_ON(buffer->transaction != NULL);
676 677
	BUG_ON(buffer->user_data < alloc->buffer);
	BUG_ON(buffer->user_data > alloc->buffer + alloc->buffer_size);
678 679

	if (buffer->async_transaction) {
680
		alloc->free_async_space += buffer_size + sizeof(struct binder_buffer);
681 682 683 684 685 686 687

		binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC_ASYNC,
			     "%d: binder_free_buf size %zd async free %zd\n",
			      alloc->pid, size, alloc->free_async_space);
	}

	binder_update_page_range(alloc, 0,
688 689 690
		(void __user *)PAGE_ALIGN((uintptr_t)buffer->user_data),
		(void __user *)(((uintptr_t)
			  buffer->user_data + buffer_size) & PAGE_MASK));
691 692 693 694

	rb_erase(&buffer->rb_node, &alloc->allocated_buffers);
	buffer->free = 1;
	if (!list_is_last(&buffer->entry, &alloc->buffers)) {
695
		struct binder_buffer *next = binder_buffer_next(buffer);
696 697 698 699 700 701 702

		if (next->free) {
			rb_erase(&next->rb_node, &alloc->free_buffers);
			binder_delete_free_buffer(alloc, next);
		}
	}
	if (alloc->buffers.next != &buffer->entry) {
703
		struct binder_buffer *prev = binder_buffer_prev(buffer);
704 705 706 707 708 709 710 711 712 713

		if (prev->free) {
			binder_delete_free_buffer(alloc, buffer);
			rb_erase(&prev->rb_node, &alloc->free_buffers);
			buffer = prev;
		}
	}
	binder_insert_free_buffer(alloc, buffer);
}

714 715
static void binder_alloc_clear_buf(struct binder_alloc *alloc,
				   struct binder_buffer *buffer);
716 717 718 719 720
/**
 * binder_alloc_free_buf() - free a binder buffer
 * @alloc:	binder_alloc for this proc
 * @buffer:	kernel pointer to buffer
 *
Y
YangHui 已提交
721
 * Free the buffer allocated via binder_alloc_new_buf()
722 723 724 725
 */
void binder_alloc_free_buf(struct binder_alloc *alloc,
			    struct binder_buffer *buffer)
{
726 727 728 729 730 731 732 733 734 735 736 737
	/*
	 * We could eliminate the call to binder_alloc_clear_buf()
	 * from binder_alloc_deferred_release() by moving this to
	 * binder_alloc_free_buf_locked(). However, that could
	 * increase contention for the alloc mutex if clear_on_free
	 * is used frequently for large buffers. The mutex is not
	 * needed for correctness here.
	 */
	if (buffer->clear_on_free) {
		binder_alloc_clear_buf(alloc, buffer);
		buffer->clear_on_free = false;
	}
738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763
	mutex_lock(&alloc->mutex);
	binder_free_buf_locked(alloc, buffer);
	mutex_unlock(&alloc->mutex);
}

/**
 * binder_alloc_mmap_handler() - map virtual address space for proc
 * @alloc:	alloc structure for this proc
 * @vma:	vma passed to mmap()
 *
 * Called by binder_mmap() to initialize the space specified in
 * vma for allocating binder buffers
 *
 * Return:
 *      0 = success
 *      -EBUSY = address space already mapped
 *      -ENOMEM = failed to map memory to given address space
 */
int binder_alloc_mmap_handler(struct binder_alloc *alloc,
			      struct vm_area_struct *vma)
{
	int ret;
	const char *failure_string;
	struct binder_buffer *buffer;

	mutex_lock(&binder_alloc_mmap_lock);
764
	if (alloc->buffer_size) {
765 766 767 768
		ret = -EBUSY;
		failure_string = "already mapped";
		goto err_already_mapped;
	}
769 770 771
	alloc->buffer_size = min_t(unsigned long, vma->vm_end - vma->vm_start,
				   SZ_4M);
	mutex_unlock(&binder_alloc_mmap_lock);
772

773
	alloc->buffer = (void __user *)vma->vm_start;
774

775
	alloc->pages = kcalloc(alloc->buffer_size / PAGE_SIZE,
K
Kees Cook 已提交
776
			       sizeof(alloc->pages[0]),
777 778 779 780 781 782 783
			       GFP_KERNEL);
	if (alloc->pages == NULL) {
		ret = -ENOMEM;
		failure_string = "alloc page array";
		goto err_alloc_pages_failed;
	}

784 785
	buffer = kzalloc(sizeof(*buffer), GFP_KERNEL);
	if (!buffer) {
786
		ret = -ENOMEM;
787 788
		failure_string = "alloc buffer struct";
		goto err_alloc_buf_struct_failed;
789
	}
790

791
	buffer->user_data = alloc->buffer;
792 793 794 795
	list_add(&buffer->entry, &alloc->buffers);
	buffer->free = 1;
	binder_insert_free_buffer(alloc, buffer);
	alloc->free_async_space = alloc->buffer_size / 2;
796
	binder_alloc_set_vma(alloc, vma);
797 798 799

	return 0;

800
err_alloc_buf_struct_failed:
801 802 803 804
	kfree(alloc->pages);
	alloc->pages = NULL;
err_alloc_pages_failed:
	alloc->buffer = NULL;
805 806
	mutex_lock(&binder_alloc_mmap_lock);
	alloc->buffer_size = 0;
807 808
err_already_mapped:
	mutex_unlock(&binder_alloc_mmap_lock);
809 810 811 812
	binder_alloc_debug(BINDER_DEBUG_USER_ERROR,
			   "%s: %d %lx-%lx %s failed %d\n", __func__,
			   alloc->pid, vma->vm_start, vma->vm_end,
			   failure_string, ret);
813 814 815 816 817 818 819 820
	return ret;
}


void binder_alloc_deferred_release(struct binder_alloc *alloc)
{
	struct rb_node *n;
	int buffers, page_count;
821
	struct binder_buffer *buffer;
822 823 824

	buffers = 0;
	mutex_lock(&alloc->mutex);
825 826
	BUG_ON(alloc->vma_addr &&
	       vma_lookup(alloc->vma_vm_mm, alloc->vma_addr));
827

828 829 830 831 832 833
	while ((n = rb_first(&alloc->allocated_buffers))) {
		buffer = rb_entry(n, struct binder_buffer, rb_node);

		/* Transaction should already have been freed */
		BUG_ON(buffer->transaction);

834 835 836 837
		if (buffer->clear_on_free) {
			binder_alloc_clear_buf(alloc, buffer);
			buffer->clear_on_free = false;
		}
838 839 840 841
		binder_free_buf_locked(alloc, buffer);
		buffers++;
	}

842 843 844 845 846 847 848 849 850 851
	while (!list_empty(&alloc->buffers)) {
		buffer = list_first_entry(&alloc->buffers,
					  struct binder_buffer, entry);
		WARN_ON(!buffer->free);

		list_del(&buffer->entry);
		WARN_ON_ONCE(!list_empty(&alloc->buffers));
		kfree(buffer);
	}

852 853 854 855 856
	page_count = 0;
	if (alloc->pages) {
		int i;

		for (i = 0; i < alloc->buffer_size / PAGE_SIZE; i++) {
857
			void __user *page_addr;
858
			bool on_lru;
859

860
			if (!alloc->pages[i].page_ptr)
861 862
				continue;

863 864
			on_lru = list_lru_del(&binder_alloc_lru,
					      &alloc->pages[i].lru);
865 866
			page_addr = alloc->buffer + i * PAGE_SIZE;
			binder_alloc_debug(BINDER_DEBUG_BUFFER_ALLOC,
867 868 869 870
				     "%s: %d: page %d at %pK %s\n",
				     __func__, alloc->pid, i, page_addr,
				     on_lru ? "on lru" : "active");
			__free_page(alloc->pages[i].page_ptr);
871 872 873 874 875
			page_count++;
		}
		kfree(alloc->pages);
	}
	mutex_unlock(&alloc->mutex);
876 877
	if (alloc->vma_vm_mm)
		mmdrop(alloc->vma_vm_mm);
878 879 880 881 882 883 884 885 886

	binder_alloc_debug(BINDER_DEBUG_OPEN_CLOSE,
		     "%s: %d buffers %d, pages %d\n",
		     __func__, alloc->pid, buffers, page_count);
}

static void print_binder_buffer(struct seq_file *m, const char *prefix,
				struct binder_buffer *buffer)
{
887
	seq_printf(m, "%s %d: %pK size %zd:%zd:%zd %s\n",
888
		   prefix, buffer->debug_id, buffer->user_data,
889
		   buffer->data_size, buffer->offsets_size,
890
		   buffer->extra_buffers_size,
891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913
		   buffer->transaction ? "active" : "delivered");
}

/**
 * binder_alloc_print_allocated() - print buffer info
 * @m:     seq_file for output via seq_printf()
 * @alloc: binder_alloc for this proc
 *
 * Prints information about every buffer associated with
 * the binder_alloc state to the given seq_file
 */
void binder_alloc_print_allocated(struct seq_file *m,
				  struct binder_alloc *alloc)
{
	struct rb_node *n;

	mutex_lock(&alloc->mutex);
	for (n = rb_first(&alloc->allocated_buffers); n != NULL; n = rb_next(n))
		print_binder_buffer(m, "  buffer",
				    rb_entry(n, struct binder_buffer, rb_node));
	mutex_unlock(&alloc->mutex);
}

914 915 916 917 918 919 920 921 922 923 924 925 926 927 928
/**
 * binder_alloc_print_pages() - print page usage
 * @m:     seq_file for output via seq_printf()
 * @alloc: binder_alloc for this proc
 */
void binder_alloc_print_pages(struct seq_file *m,
			      struct binder_alloc *alloc)
{
	struct binder_lru_page *page;
	int i;
	int active = 0;
	int lru = 0;
	int free = 0;

	mutex_lock(&alloc->mutex);
929 930 931 932
	/*
	 * Make sure the binder_alloc is fully initialized, otherwise we might
	 * read inconsistent state.
	 */
933 934 935 936 937

	mmap_read_lock(alloc->vma_vm_mm);
	if (binder_alloc_get_vma(alloc) == NULL) {
		mmap_read_unlock(alloc->vma_vm_mm);
		goto uninitialized;
938
	}
939 940 941 942 943 944 945 946 947 948 949 950 951

	mmap_read_unlock(alloc->vma_vm_mm);
	for (i = 0; i < alloc->buffer_size / PAGE_SIZE; i++) {
		page = &alloc->pages[i];
		if (!page->page_ptr)
			free++;
		else if (list_empty(&page->lru))
			active++;
		else
			lru++;
	}

uninitialized:
952 953
	mutex_unlock(&alloc->mutex);
	seq_printf(m, "  pages: %d:%d:%d\n", active, lru, free);
954
	seq_printf(m, "  pages high watermark: %zu\n", alloc->pages_high);
955 956
}

957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985
/**
 * binder_alloc_get_allocated_count() - return count of buffers
 * @alloc: binder_alloc for this proc
 *
 * Return: count of allocated buffers
 */
int binder_alloc_get_allocated_count(struct binder_alloc *alloc)
{
	struct rb_node *n;
	int count = 0;

	mutex_lock(&alloc->mutex);
	for (n = rb_first(&alloc->allocated_buffers); n != NULL; n = rb_next(n))
		count++;
	mutex_unlock(&alloc->mutex);
	return count;
}


/**
 * binder_alloc_vma_close() - invalidate address space
 * @alloc: binder_alloc for this proc
 *
 * Called from binder_vma_close() when releasing address space.
 * Clears alloc->vma to prevent new incoming transactions from
 * allocating more buffers.
 */
void binder_alloc_vma_close(struct binder_alloc *alloc)
{
986
	binder_alloc_set_vma(alloc, NULL);
987 988
}

989 990 991 992 993 994 995 996 997 998 999 1000 1001
/**
 * binder_alloc_free_page() - shrinker callback to free pages
 * @item:   item to free
 * @lock:   lock protecting the item
 * @cb_arg: callback argument
 *
 * Called from list_lru_walk() in binder_shrink_scan() to free
 * up pages when the system is under memory pressure.
 */
enum lru_status binder_alloc_free_page(struct list_head *item,
				       struct list_lru_one *lru,
				       spinlock_t *lock,
				       void *cb_arg)
1002
	__must_hold(lock)
1003 1004 1005 1006 1007 1008 1009 1010
{
	struct mm_struct *mm = NULL;
	struct binder_lru_page *page = container_of(item,
						    struct binder_lru_page,
						    lru);
	struct binder_alloc *alloc;
	uintptr_t page_addr;
	size_t index;
1011
	struct vm_area_struct *vma;
1012 1013 1014 1015 1016 1017 1018 1019 1020 1021

	alloc = page->alloc;
	if (!mutex_trylock(&alloc->mutex))
		goto err_get_alloc_mutex_failed;

	if (!page->page_ptr)
		goto err_page_already_freed;

	index = page - alloc->pages;
	page_addr = (uintptr_t)alloc->buffer + index * PAGE_SIZE;
1022 1023 1024 1025

	mm = alloc->vma_vm_mm;
	if (!mmget_not_zero(mm))
		goto err_mmget;
1026
	if (!mmap_read_trylock(mm))
1027
		goto err_mmap_read_lock_failed;
1028
	vma = binder_alloc_get_vma(alloc);
1029 1030 1031

	list_lru_isolate(lru, item);
	spin_unlock(lock);
1032

1033
	if (vma) {
1034 1035
		trace_binder_unmap_user_start(alloc, index);

T
Todd Kjos 已提交
1036
		zap_page_range(vma, page_addr, PAGE_SIZE);
1037

1038
		trace_binder_unmap_user_end(alloc, index);
1039
	}
1040
	mmap_read_unlock(mm);
1041
	mmput_async(mm);
1042

1043 1044
	trace_binder_unmap_kernel_start(alloc, index);

1045 1046 1047
	__free_page(page->page_ptr);
	page->page_ptr = NULL;

1048 1049
	trace_binder_unmap_kernel_end(alloc, index);

1050
	spin_lock(lock);
1051
	mutex_unlock(&alloc->mutex);
1052
	return LRU_REMOVED_RETRY;
1053

1054
err_mmap_read_lock_failed:
1055
	mmput_async(mm);
1056
err_mmget:
1057 1058 1059 1060 1061 1062 1063 1064 1065
err_page_already_freed:
	mutex_unlock(&alloc->mutex);
err_get_alloc_mutex_failed:
	return LRU_SKIP;
}

static unsigned long
binder_shrink_count(struct shrinker *shrink, struct shrink_control *sc)
{
1066
	return list_lru_count(&binder_alloc_lru);
1067 1068 1069 1070 1071
}

static unsigned long
binder_shrink_scan(struct shrinker *shrink, struct shrink_control *sc)
{
1072
	return list_lru_walk(&binder_alloc_lru, binder_alloc_free_page,
1073 1074 1075
			    NULL, sc->nr_to_scan);
}

1076
static struct shrinker binder_shrinker = {
1077 1078 1079 1080 1081
	.count_objects = binder_shrink_count,
	.scan_objects = binder_shrink_scan,
	.seeks = DEFAULT_SEEKS,
};

1082 1083 1084 1085 1086 1087 1088 1089 1090 1091
/**
 * binder_alloc_init() - called by binder_open() for per-proc initialization
 * @alloc: binder_alloc for this proc
 *
 * Called from binder_open() to initialize binder_alloc fields for
 * new binder proc
 */
void binder_alloc_init(struct binder_alloc *alloc)
{
	alloc->pid = current->group_leader->pid;
1092 1093
	alloc->vma_vm_mm = current->mm;
	mmgrab(alloc->vma_vm_mm);
1094
	mutex_init(&alloc->mutex);
1095
	INIT_LIST_HEAD(&alloc->buffers);
1096 1097
}

1098
int binder_alloc_shrinker_init(void)
1099
{
1100 1101 1102
	int ret = list_lru_init(&binder_alloc_lru);

	if (ret == 0) {
1103
		ret = register_shrinker(&binder_shrinker, "android-binder");
1104 1105 1106 1107
		if (ret)
			list_lru_destroy(&binder_alloc_lru);
	}
	return ret;
1108
}
1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149

/**
 * check_buffer() - verify that buffer/offset is safe to access
 * @alloc: binder_alloc for this proc
 * @buffer: binder buffer to be accessed
 * @offset: offset into @buffer data
 * @bytes: bytes to access from offset
 *
 * Check that the @offset/@bytes are within the size of the given
 * @buffer and that the buffer is currently active and not freeable.
 * Offsets must also be multiples of sizeof(u32). The kernel is
 * allowed to touch the buffer in two cases:
 *
 * 1) when the buffer is being created:
 *     (buffer->free == 0 && buffer->allow_user_free == 0)
 * 2) when the buffer is being torn down:
 *     (buffer->free == 0 && buffer->transaction == NULL).
 *
 * Return: true if the buffer is safe to access
 */
static inline bool check_buffer(struct binder_alloc *alloc,
				struct binder_buffer *buffer,
				binder_size_t offset, size_t bytes)
{
	size_t buffer_size = binder_alloc_buffer_size(alloc, buffer);

	return buffer_size >= bytes &&
		offset <= buffer_size - bytes &&
		IS_ALIGNED(offset, sizeof(u32)) &&
		!buffer->free &&
		(!buffer->allow_user_free || !buffer->transaction);
}

/**
 * binder_alloc_get_page() - get kernel pointer for given buffer offset
 * @alloc: binder_alloc for this proc
 * @buffer: binder buffer to be accessed
 * @buffer_offset: offset into @buffer data
 * @pgoffp: address to copy final page offset to
 *
 * Lookup the struct page corresponding to the address
1150
 * at @buffer_offset into @buffer->user_data. If @pgoffp is not
1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166
 * NULL, the byte-offset into the page is written there.
 *
 * The caller is responsible to ensure that the offset points
 * to a valid address within the @buffer and that @buffer is
 * not freeable by the user. Since it can't be freed, we are
 * guaranteed that the corresponding elements of @alloc->pages[]
 * cannot change.
 *
 * Return: struct page
 */
static struct page *binder_alloc_get_page(struct binder_alloc *alloc,
					  struct binder_buffer *buffer,
					  binder_size_t buffer_offset,
					  pgoff_t *pgoffp)
{
	binder_size_t buffer_space_offset = buffer_offset +
1167
		(buffer->user_data - alloc->buffer);
1168 1169 1170 1171 1172 1173 1174 1175 1176
	pgoff_t pgoff = buffer_space_offset & ~PAGE_MASK;
	size_t index = buffer_space_offset >> PAGE_SHIFT;
	struct binder_lru_page *lru_page;

	lru_page = &alloc->pages[index];
	*pgoffp = pgoff;
	return lru_page->page_ptr;
}

1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197
/**
 * binder_alloc_clear_buf() - zero out buffer
 * @alloc: binder_alloc for this proc
 * @buffer: binder buffer to be cleared
 *
 * memset the given buffer to 0
 */
static void binder_alloc_clear_buf(struct binder_alloc *alloc,
				   struct binder_buffer *buffer)
{
	size_t bytes = binder_alloc_buffer_size(alloc, buffer);
	binder_size_t buffer_offset = 0;

	while (bytes) {
		unsigned long size;
		struct page *page;
		pgoff_t pgoff;

		page = binder_alloc_get_page(alloc, buffer,
					     buffer_offset, &pgoff);
		size = min_t(size_t, bytes, PAGE_SIZE - pgoff);
1198
		memset_page(page, pgoff, 0, size);
1199 1200 1201 1202 1203
		bytes -= size;
		buffer_offset += size;
	}
}

1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235
/**
 * binder_alloc_copy_user_to_buffer() - copy src user to tgt user
 * @alloc: binder_alloc for this proc
 * @buffer: binder buffer to be accessed
 * @buffer_offset: offset into @buffer data
 * @from: userspace pointer to source buffer
 * @bytes: bytes to copy
 *
 * Copy bytes from source userspace to target buffer.
 *
 * Return: bytes remaining to be copied
 */
unsigned long
binder_alloc_copy_user_to_buffer(struct binder_alloc *alloc,
				 struct binder_buffer *buffer,
				 binder_size_t buffer_offset,
				 const void __user *from,
				 size_t bytes)
{
	if (!check_buffer(alloc, buffer, buffer_offset, bytes))
		return bytes;

	while (bytes) {
		unsigned long size;
		unsigned long ret;
		struct page *page;
		pgoff_t pgoff;
		void *kptr;

		page = binder_alloc_get_page(alloc, buffer,
					     buffer_offset, &pgoff);
		size = min_t(size_t, bytes, PAGE_SIZE - pgoff);
1236
		kptr = kmap_local_page(page) + pgoff;
1237
		ret = copy_from_user(kptr, from, size);
1238
		kunmap_local(kptr);
1239 1240 1241 1242 1243 1244 1245 1246
		if (ret)
			return bytes - size + ret;
		bytes -= size;
		from += size;
		buffer_offset += size;
	}
	return 0;
}
1247

1248 1249 1250 1251 1252 1253
static int binder_alloc_do_buffer_copy(struct binder_alloc *alloc,
				       bool to_buffer,
				       struct binder_buffer *buffer,
				       binder_size_t buffer_offset,
				       void *ptr,
				       size_t bytes)
1254 1255
{
	/* All copies must be 32-bit aligned and 32-bit size */
1256 1257
	if (!check_buffer(alloc, buffer, buffer_offset, bytes))
		return -EINVAL;
1258 1259 1260 1261 1262 1263 1264 1265 1266 1267

	while (bytes) {
		unsigned long size;
		struct page *page;
		pgoff_t pgoff;

		page = binder_alloc_get_page(alloc, buffer,
					     buffer_offset, &pgoff);
		size = min_t(size_t, bytes, PAGE_SIZE - pgoff);
		if (to_buffer)
1268
			memcpy_to_page(page, pgoff, ptr, size);
1269
		else
1270
			memcpy_from_page(ptr, page, pgoff, size);
1271 1272 1273 1274 1275
		bytes -= size;
		pgoff = 0;
		ptr = ptr + size;
		buffer_offset += size;
	}
1276
	return 0;
1277 1278
}

1279 1280 1281 1282 1283
int binder_alloc_copy_to_buffer(struct binder_alloc *alloc,
				struct binder_buffer *buffer,
				binder_size_t buffer_offset,
				void *src,
				size_t bytes)
1284
{
1285 1286
	return binder_alloc_do_buffer_copy(alloc, true, buffer, buffer_offset,
					   src, bytes);
1287 1288
}

1289 1290 1291 1292 1293
int binder_alloc_copy_from_buffer(struct binder_alloc *alloc,
				  void *dest,
				  struct binder_buffer *buffer,
				  binder_size_t buffer_offset,
				  size_t bytes)
1294
{
1295 1296
	return binder_alloc_do_buffer_copy(alloc, false, buffer, buffer_offset,
					   dest, bytes);
1297 1298
}