Searched refs:last_alloc (Results 1 – 5 of 5) sorted by relevance
/linux/mm/ |
H A D | percpu-stats.c | 56 int i, last_alloc, as_len, start, end; in chunk_map_stats() local 69 last_alloc = find_last_bit(chunk->alloc_map, in chunk_map_stats() 72 last_alloc = test_bit(last_alloc, chunk->alloc_map) ? in chunk_map_stats() 73 last_alloc + 1 : 0; in chunk_map_stats() 87 while (start < last_alloc) { in chunk_map_stats() 89 end = find_next_bit(chunk->bound_map, last_alloc, in chunk_map_stats() 93 end = find_next_bit(chunk->alloc_map, last_alloc, in chunk_map_stats()
|
/linux/fs/nilfs2/ |
H A D | sufile.c | 322 __u64 segnum, maxsegnum, last_alloc; in nilfs_sufile_alloc() local 334 last_alloc = le64_to_cpu(header->sh_last_alloc); in nilfs_sufile_alloc() 339 segnum = last_alloc + 1; in nilfs_sufile_alloc() 352 maxsegnum = last_alloc; in nilfs_sufile_alloc()
|
/linux/fs/btrfs/ |
H A D | zoned.c | 1556 u64 last_alloc = 0; in btrfs_load_block_group_zone_info() local 1606 ret = calculate_alloc_pointer(cache, &last_alloc, new); in btrfs_load_block_group_zone_info() 1613 cache->alloc_offset = last_alloc; in btrfs_load_block_group_zone_info() 1683 if (!ret && num_conventional && last_alloc > cache->alloc_offset) { in btrfs_load_block_group_zone_info() 1686 logical, last_alloc, cache->alloc_offset); in btrfs_load_block_group_zone_info()
|
H A D | inode.c | 8764 u64 last_alloc = (u64)-1; in __btrfs_prealloc_file_range() local 8780 cur_bytes = min(cur_bytes, last_alloc); in __btrfs_prealloc_file_range() 8795 last_alloc = ins.offset; in __btrfs_prealloc_file_range()
|
/linux/tools/perf/ |
H A D | builtin-kmem.c | 66 u64 last_alloc; member 127 data->last_alloc = bytes_alloc; in insert_alloc_stat() 255 total_freed += s_alloc->last_alloc; in evsel__process_free_event()
|