Lines Matching full:buffer_size
65 return alloc->vm_start + alloc->buffer_size - buffer->user_data; in binder_alloc_buffer_size()
76 size_t buffer_size; in binder_insert_free_buffer() local
92 buffer_size = binder_alloc_buffer_size(alloc, buffer); in binder_insert_free_buffer()
94 if (new_buffer_size < buffer_size) in binder_insert_free_buffer()
434 size_t buffer_size; in debug_no_space_locked() local
439 buffer_size = binder_alloc_buffer_size(alloc, buffer); in debug_no_space_locked()
441 total_alloc_size += buffer_size; in debug_no_space_locked()
442 if (buffer_size > largest_alloc_size) in debug_no_space_locked()
443 largest_alloc_size = buffer_size; in debug_no_space_locked()
448 buffer_size = binder_alloc_buffer_size(alloc, buffer); in debug_no_space_locked()
450 total_free_size += buffer_size; in debug_no_space_locked()
451 if (buffer_size > largest_free_size) in debug_no_space_locked()
452 largest_free_size = buffer_size; in debug_no_space_locked()
481 if (alloc->free_async_space >= alloc->buffer_size / 10) { in debug_low_async_space_locked()
502 if (num_buffers > 50 || total_alloc_size > alloc->buffer_size / 4) { in debug_low_async_space_locked()
526 size_t buffer_size; in binder_alloc_new_buf_locked() local
539 buffer_size = binder_alloc_buffer_size(alloc, buffer); in binder_alloc_new_buf_locked()
541 if (size < buffer_size) { in binder_alloc_new_buf_locked()
544 } else if (size > buffer_size) { in binder_alloc_new_buf_locked()
561 if (buffer_size != size) { in binder_alloc_new_buf_locked()
564 buffer_size = binder_alloc_buffer_size(alloc, buffer); in binder_alloc_new_buf_locked()
566 WARN_ON(n || buffer_size == size); in binder_alloc_new_buf_locked()
576 alloc->pid, size, buffer, buffer_size); in binder_alloc_new_buf_locked()
580 * with buffer_size determines if the last page is shared with an in binder_alloc_new_buf_locked()
584 next_used_page = (buffer->user_data + buffer_size) & PAGE_MASK; in binder_alloc_new_buf_locked()
742 size_t size, buffer_size; in binder_free_buf_locked() local
744 buffer_size = binder_alloc_buffer_size(alloc, buffer); in binder_free_buf_locked()
751 "%d: binder_free_buf %pK size %zd buffer_size %zd\n", in binder_free_buf_locked()
752 alloc->pid, buffer, size, buffer_size); in binder_free_buf_locked()
755 BUG_ON(size > buffer_size); in binder_free_buf_locked()
758 BUG_ON(buffer->user_data > alloc->vm_start + alloc->buffer_size); in binder_free_buf_locked()
761 alloc->free_async_space += buffer_size; in binder_free_buf_locked()
768 (buffer->user_data + buffer_size) & PAGE_MASK); in binder_free_buf_locked()
908 if (alloc->buffer_size) { in binder_alloc_mmap_handler()
913 alloc->buffer_size = min_t(unsigned long, vma->vm_end - vma->vm_start, in binder_alloc_mmap_handler()
919 alloc->pages = kvcalloc(alloc->buffer_size / PAGE_SIZE, in binder_alloc_mmap_handler()
939 alloc->free_async_space = alloc->buffer_size / 2; in binder_alloc_mmap_handler()
952 alloc->buffer_size = 0; in binder_alloc_mmap_handler()
1002 for (i = 0; i < alloc->buffer_size / PAGE_SIZE; i++) { in binder_alloc_deferred_release()
1079 for (i = 0; i < alloc->buffer_size / PAGE_SIZE; i++) { in binder_alloc_print_pages()
1308 size_t buffer_size = binder_alloc_buffer_size(alloc, buffer); in check_buffer() local
1310 return buffer_size >= bytes && in check_buffer()
1311 offset <= buffer_size - bytes && in check_buffer()