| /linux/arch/arm/mm/ |
| H A D | cache-feroceon-l2.c | 141 unsigned long range_end; in calc_range_end() local 149 range_end = end; in calc_range_end() 156 if (range_end > start + MAX_RANGE_SIZE) in calc_range_end() 157 range_end = start + MAX_RANGE_SIZE; in calc_range_end() 162 if (range_end > (start | (PAGE_SIZE - 1)) + 1) in calc_range_end() 163 range_end = (start | (PAGE_SIZE - 1)) + 1; in calc_range_end() 165 return range_end; in calc_range_end() 190 unsigned long range_end = calc_range_end(start, end); in feroceon_l2_inv_range() local 191 l2_inv_pa_range(start, range_end - CACHE_LINE_SIZE); in feroceon_l2_inv_range() 192 start = range_end; in feroceon_l2_inv_range() [all …]
|
| H A D | pageattr.c | 29 unsigned long range_start, unsigned long range_end) in range_in_range() argument 31 return start >= range_start && start < range_end && in range_in_range() 32 size <= range_end - start; in range_in_range()
|
| /linux/tools/testing/selftests/kvm/include/ |
| H A D | sparsebit.h | 82 #define sparsebit_for_each_set_range(s, range_begin, range_end) \ argument 84 range_end = sparsebit_next_clear(s, range_begin) - 1; \ 85 range_begin && range_end; \ 86 range_begin = sparsebit_next_set(s, range_end), \ 87 range_end = sparsebit_next_clear(s, range_begin) - 1)
|
| /linux/fs/btrfs/ |
| H A D | fiemap.c | 179 const u64 range_end = offset + len; in emit_fiemap_extent() local 209 if (range_end <= cache_end) in emit_fiemap_extent() 216 len = range_end - cache_end; in emit_fiemap_extent() 643 u64 range_end; in extent_fiemap() local 661 range_end = round_up(start + len, sectorsize); in extent_fiemap() 664 btrfs_lock_extent(&inode->io_tree, range_start, range_end, &cached_state); in extent_fiemap() 684 while (prev_extent_end < range_end) { in extent_fiemap() 714 const u64 hole_end = min(key.offset, range_end) - 1; in extent_fiemap() 729 if (key.offset >= range_end) { in extent_fiemap() 813 if (!stopped && prev_extent_end < range_end) { in extent_fiemap() [all …]
|
| /linux/drivers/firmware/efi/ |
| H A D | unaccepted_memory.c | 36 unsigned long range_start, range_end; in accept_memory() local 127 for_each_set_bitrange_from(range_start, range_end, unaccepted->bitmap, in accept_memory() 130 unsigned long len = range_end - range_start; in accept_memory() 133 phys_end = range_end * unit_size + unaccepted->phys_base; in accept_memory()
|
| /linux/net/bridge/ |
| H A D | br_vlan_options.c | 33 const struct net_bridge_vlan *range_end) in __vlan_tun_can_enter_range() argument 35 return (!v_curr->tinfo.tunnel_dst && !range_end->tinfo.tunnel_dst) || in __vlan_tun_can_enter_range() 36 vlan_tunid_inrange(v_curr, range_end); in __vlan_tun_can_enter_range() 41 const struct net_bridge_vlan *range_end) in br_vlan_opts_eq_range() argument 43 u8 range_mc_rtr = br_vlan_multicast_router(range_end); in br_vlan_opts_eq_range() 46 return v_curr->state == range_end->state && in br_vlan_opts_eq_range() 47 __vlan_tun_can_enter_range(v_curr, range_end) && in br_vlan_opts_eq_range() 266 struct net_bridge_vlan *range_end, in br_vlan_process_options() argument 284 if (!range_end || !br_vlan_should_use(range_end)) { in br_vlan_process_options() 290 for (vid = range_start->vid; vid <= range_end->vid; vid++) { in br_vlan_process_options()
|
| H A D | br_vlan.c | 1980 const struct net_bridge_vlan *range_end) in br_vlan_can_enter_range() argument 1982 return v_curr->vid - range_end->vid == 1 && in br_vlan_can_enter_range() 1983 range_end->flags == v_curr->flags && in br_vlan_can_enter_range() 1984 br_vlan_opts_eq_range(v_curr, range_end); in br_vlan_can_enter_range() 1992 struct net_bridge_vlan *v, *range_start = NULL, *range_end = NULL; in br_vlan_dump_dev() local 2047 range_end = v; in br_vlan_dump_dev() 2052 if (br_vlan_global_opts_can_enter_range(v, range_end)) in br_vlan_dump_dev() 2055 range_end->vid, in br_vlan_dump_dev() 2061 idx += range_end->vid - range_start->vid + 1; in br_vlan_dump_dev() 2065 !br_vlan_can_enter_range(v, range_end)) { in br_vlan_dump_dev() [all …]
|
| /linux/drivers/firmware/efi/libstub/ |
| H A D | unaccepted_memory.c | 182 unsigned long range_start, range_end; in accept_memory() local 212 for_each_set_bitrange_from(range_start, range_end, in accept_memory() 217 phys_end = range_end * unit_size + unaccepted_table->phys_base; in accept_memory() 221 range_start, range_end - range_start); in accept_memory()
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | fault.c | 111 unsigned long range_start, range_end, i; in fault_opcodes_write() local 129 if (kstrtoul(token, 0, &range_end)) in fault_opcodes_write() 132 range_end = range_start; in fault_opcodes_write() 134 if (range_start == range_end && range_start == -1UL) { in fault_opcodes_write() 140 if (range_start >= bound || range_end >= bound) in fault_opcodes_write() 143 for (i = range_start; i <= range_end; i++) { in fault_opcodes_write()
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_doorbell.c | 213 int range_end = dev->shared_resources.non_cp_doorbells_end; in init_doorbell_bitmap() local 219 pr_debug("reserved doorbell 0x%03x - 0x%03x\n", range_start, range_end); in init_doorbell_bitmap() 222 range_end + KFD_QUEUE_DOORBELL_MIRROR_OFFSET); in init_doorbell_bitmap() 225 if (i >= range_start && i <= range_end) { in init_doorbell_bitmap()
|
| H A D | kfd_device.c | 1333 (*mem_obj)->range_end = found; in kfd_gtt_sa_allocate() 1356 (*mem_obj)->range_end = in kfd_gtt_sa_allocate() 1364 if ((*mem_obj)->range_end != found) { in kfd_gtt_sa_allocate() 1384 (*mem_obj)->range_start, (*mem_obj)->range_end); in kfd_gtt_sa_allocate() 1388 (*mem_obj)->range_end - (*mem_obj)->range_start + 1); in kfd_gtt_sa_allocate() 1410 mem_obj, mem_obj->range_start, mem_obj->range_end); in kfd_gtt_sa_free() 1416 mem_obj->range_end - mem_obj->range_start + 1); in kfd_gtt_sa_free()
|
| /linux/arch/powerpc/platforms/powernv/ |
| H A D | opal-prd.c | 64 uint64_t range_addr, range_size, range_end; in opal_prd_range_is_valid() local 73 range_end = range_addr + range_size; in opal_prd_range_is_valid() 81 if (range_end <= range_addr) in opal_prd_range_is_valid() 84 if (addr >= range_addr && addr + size <= range_end) { in opal_prd_range_is_valid()
|
| /linux/arch/s390/boot/ |
| H A D | physmem_info.c | 278 unsigned long range_start, range_end; in __physmem_alloc_range() local 284 __get_physmem_range(nranges - 1, &range_start, &range_end, false); in __physmem_alloc_range() 285 pos = min(range_end, pos); in __physmem_alloc_range()
|
| /linux/fs/ocfs2/ |
| H A D | file.c | 861 u64 *range_start, u64 *range_end) in ocfs2_zero_extend_get_range() argument 889 *range_end = 0; in ocfs2_zero_extend_get_range() 921 *range_end = ocfs2_clusters_to_bytes(inode->i_sb, in ocfs2_zero_extend_get_range() 933 u64 range_end, struct buffer_head *di_bh) in ocfs2_zero_extend_range() argument 942 (unsigned long long)range_end); in ocfs2_zero_extend_range() 943 BUG_ON(range_start >= range_end); in ocfs2_zero_extend_range() 945 while (zero_pos < range_end) { in ocfs2_zero_extend_range() 947 if (next_pos > range_end) in ocfs2_zero_extend_range() 948 next_pos = range_end; in ocfs2_zero_extend_range() 970 u64 zero_start, range_start = 0, range_end = 0; in ocfs2_zero_extend() local [all …]
|
| /linux/drivers/gpu/drm/msm/ |
| H A D | msm_gem.c | 466 u64 range_end) in get_vma_locked() argument 475 vma = msm_gem_vma_new(vm, obj, 0, range_start, range_end); in get_vma_locked() 478 GEM_WARN_ON((vma->va.addr + obj->size) > range_end); in get_vma_locked() 553 u64 range_start, u64 range_end) in get_and_pin_iova_range_locked() argument 563 vma = get_vma_locked(obj, vm, range_start, range_end); in get_and_pin_iova_range_locked() 582 u64 range_start, u64 range_end) in msm_gem_get_and_pin_iova_range() argument 588 ret = get_and_pin_iova_range_locked(obj, vm, iova, range_start, range_end); in msm_gem_get_and_pin_iova_range()
|
| H A D | msm_gem.h | 190 u64 offset, u64 range_start, u64 range_end); 278 u64 range_start, u64 range_end);
|
| H A D | msm_gem_vma.c | 369 u64 offset, u64 range_start, u64 range_end) in msm_gem_vma_new() argument 387 range_start, range_end, 0); in msm_gem_vma_new() 393 range_end = range_start + obj->size; in msm_gem_vma_new() 397 GEM_WARN_ON((range_end - range_start) > obj->size); in msm_gem_vma_new() 401 .va.range = range_end - range_start, in msm_gem_vma_new()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_svm.c | 288 static int xe_svm_range_set_default_attr(struct xe_vm *vm, u64 range_start, u64 range_end) in xe_svm_range_set_default_attr() argument 315 if (xe_vma_start(vma) == range_start && xe_vma_end(vma) == range_end) { in xe_svm_range_set_default_attr() 321 range_start, range_end); in xe_svm_range_set_default_attr() 322 err = xe_vm_alloc_cpu_addr_mirror_vma(vm, range_start, range_end - range_start); in xe_svm_range_set_default_attr() 341 u64 range_end; in xe_svm_garbage_collector() local 358 range_end = xe_svm_range_end(range); in xe_svm_garbage_collector() 372 err = xe_svm_range_set_default_attr(vm, range_start, range_end); in xe_svm_garbage_collector()
|
| /linux/include/trace/events/ |
| H A D | writeback.h | 493 __field(long, range_end) 506 __entry->range_end = (long)wbc->range_end; 520 __entry->range_end,
|
| /linux/drivers/accel/habanalabs/goya/ |
| H A D | goya_coresight.c | 371 u64 range_start, range_end; in goya_etr_validate_address() local 380 range_end = prop->dmmu.end_addr; in goya_etr_validate_address() 382 return hl_mem_area_inside_range(addr, size, range_start, range_end); in goya_etr_validate_address()
|
| /linux/drivers/net/ethernet/netronome/nfp/bpf/ |
| H A D | jit.c | 2734 s16 range_end = meta->pkt_cache.range_end; in mem_ldx_data_init_pktcache() local 2741 len = range_end - range_start; in mem_ldx_data_init_pktcache() 2866 if (meta->pkt_cache.range_end) { in mem_ldx() 4279 s16 range_start = 0, range_end = 0; in nfp_bpf_opt_pkt_cache() local 4332 s16 new_end = range_end; in nfp_bpf_opt_pkt_cache() 4341 if (end > range_end) { in nfp_bpf_opt_pkt_cache() 4352 range_end = new_end; in nfp_bpf_opt_pkt_cache() 4359 range_node->pkt_cache.range_end = range_end; in nfp_bpf_opt_pkt_cache() 4366 range_end = insn->off + BPF_LDST_BYTES(insn); in nfp_bpf_opt_pkt_cache() 4371 range_node->pkt_cache.range_end = range_end; in nfp_bpf_opt_pkt_cache() [all …]
|
| /linux/fs/nfs/ |
| H A D | nfstrace.h | 301 loff_t range_end 304 TP_ARGS(inode, range_start, range_end), 312 __field(loff_t, range_end) 323 __entry->range_end = range_end; 332 __entry->range_start, __entry->range_end 341 loff_t range_end \ 343 TP_ARGS(inode, range_start, range_end))
|
| /linux/fs/iomap/ |
| H A D | buffered-io.c | 128 struct iomap_folio_state *ifs, u64 *range_start, u64 range_end) in ifs_find_dirty_range() argument 134 offset_in_folio(folio, range_end) >> inode->i_blkbits, in ifs_find_dirty_range() 152 u64 range_end) in iomap_find_dirty_range() argument 156 if (*range_start >= range_end) in iomap_find_dirty_range() 160 return ifs_find_dirty_range(folio, ifs, range_start, range_end); in iomap_find_dirty_range() 161 return range_end - *range_start; in iomap_find_dirty_range()
|
| /linux/arch/powerpc/kernel/ |
| H A D | fadump.c | 1748 phys_addr_t range_start, range_end; in fadump_setup_param_area() local 1763 range_end = memblock_end_of_DRAM(); in fadump_setup_param_area() 1778 range_end = min(ppc64_rma_size, fw_dump.boot_mem_top); in fadump_setup_param_area() 1784 range_end); in fadump_setup_param_area()
|
| /linux/include/linux/ |
| H A D | writeback.h | 55 loff_t range_end; member
|