| /linux/kernel/ |
| H A D | range.c | 12 int add_range(struct range *range, int az, int nr_range, u64 start, u64 end) in add_range() argument 21 range[nr_range].start = start; in add_range() 22 range[nr_range].end = end; in add_range() 29 int add_range_with_merge(struct range *range, int az, int nr_range, in add_range_with_merge() argument 41 if (!range[i].end) in add_range_with_merge() 44 common_start = max(range[i].start, start); in add_range_with_merge() 45 common_end = min(range[i].end, end); in add_range_with_merge() 50 start = min(range[i].start, start); in add_range_with_merge() 51 end = max(range[i].end, end); in add_range_with_merge() 53 memmove(&range[i], &range[i + 1], in add_range_with_merge() [all …]
|
| /linux/drivers/iommu/generic_pt/ |
| H A D | pt_iter.h | 28 static inline int pt_check_range(struct pt_range *range) in pt_check_range() argument 32 PT_WARN_ON(!range->max_vasz_lg2); in pt_check_range() 34 if (pt_feature(range->common, PT_FEAT_SIGN_EXTEND)) { in pt_check_range() 35 PT_WARN_ON(range->common->max_vasz_lg2 != range->max_vasz_lg2); in pt_check_range() 36 prefix = fvalog2_div(range->va, range->max_vasz_lg2 - 1) ? in pt_check_range() 40 prefix = pt_full_va_prefix(range->common); in pt_check_range() 43 if (!fvalog2_div_eq(range->va, prefix, range->max_vasz_lg2) || in pt_check_range() 44 !fvalog2_div_eq(range->last_va, prefix, range->max_vasz_lg2)) in pt_check_range() 61 pts->range->va = fvalog2_set_mod(pts->range->va, lower_va, in pt_index_to_va() 88 struct pt_range *range = pts->range; in pt_entry_fully_covered() local [all …]
|
| /linux/drivers/soc/ti/ |
| H A D | knav_qmss_acc.c | 20 #define knav_range_offset_to_inst(kdev, range, q) \ argument 21 (range->queue_base_inst + (q << kdev->inst_shift)) 23 static void __knav_acc_notify(struct knav_range_info *range, in __knav_acc_notify() argument 26 struct knav_device *kdev = range->kdev; in __knav_acc_notify() 30 range_base = kdev->base_id + range->queue_base; in __knav_acc_notify() 32 if (range->flags & RANGE_MULTI_QUEUE) { in __knav_acc_notify() 33 for (queue = 0; queue < range->num_queues; queue++) { in __knav_acc_notify() 34 inst = knav_range_offset_to_inst(kdev, range, in __knav_acc_notify() 44 queue = acc->channel - range->acc_info.start_channel; in __knav_acc_notify() 45 inst = knav_range_offset_to_inst(kdev, range, queue); in __knav_acc_notify() [all …]
|
| H A D | knav_qmss_queue.c | 114 static int knav_queue_setup_irq(struct knav_range_info *range, in knav_queue_setup_irq() argument 117 unsigned queue = inst->id - range->queue_base; in knav_queue_setup_irq() 120 if (range->flags & RANGE_HAS_IRQ) { in knav_queue_setup_irq() 121 irq = range->irqs[queue].irq; in knav_queue_setup_irq() 126 if (range->irqs[queue].cpu_mask) { in knav_queue_setup_irq() 127 ret = irq_set_affinity_hint(irq, range->irqs[queue].cpu_mask); in knav_queue_setup_irq() 129 dev_warn(range->kdev->dev, in knav_queue_setup_irq() 140 struct knav_range_info *range = inst->range; in knav_queue_free_irq() local 141 unsigned queue = inst->id - inst->range->queue_base; in knav_queue_free_irq() 144 if (range->flags & RANGE_HAS_IRQ) { in knav_queue_free_irq() [all …]
|
| /linux/arch/s390/include/asm/ |
| H A D | physmem_info.h | 138 #define for_each_physmem_reserved_type_range(t, range, p_start, p_end) \ argument 139 for (range = &physmem_info.reserved[t], *p_start = range->start, *p_end = range->end; \ 140 range && range->end; range = range->chain ? __va(range->chain) : NULL, \ 141 *p_start = range ? range->start : 0, *p_end = range ? range->end : 0) 144 struct reserved_range *range) in __physmem_reserved_next() argument 146 if (!range) { in __physmem_reserved_next() 147 range = &physmem_info.reserved[*t]; in __physmem_reserved_next() 148 if (range->end) in __physmem_reserved_next() 149 return range; in __physmem_reserved_next() 151 if (range->chain) in __physmem_reserved_next() [all …]
|
| /linux/mm/ |
| H A D | memremap.c | 41 static void pgmap_array_delete(struct range *range) in pgmap_array_delete() argument 43 xa_store_range(&pgmap_array, PHYS_PFN(range->start), PHYS_PFN(range->end), in pgmap_array_delete() 50 struct range *range = &pgmap->ranges[range_id]; in pfn_first() local 51 unsigned long pfn = PHYS_PFN(range->start); in pfn_first() 63 struct range *range = &pgmap->ranges[i]; in pgmap_pfn_valid() local 65 if (pfn >= PHYS_PFN(range->start) && in pgmap_pfn_valid() 66 pfn <= PHYS_PFN(range->end)) in pgmap_pfn_valid() 75 const struct range *range = &pgmap->ranges[range_id]; in pfn_end() local 77 return (range->start + range_len(range)) >> PAGE_SHIFT; in pfn_end() 88 struct range *range = &pgmap->ranges[range_id]; in pageunmap_range() local [all …]
|
| H A D | hmm.c | 34 struct hmm_range *range; member 51 struct hmm_range *range, unsigned long cpu_flags) in hmm_pfns_fill() argument 53 unsigned long i = (addr - range->start) >> PAGE_SHIFT; in hmm_pfns_fill() 56 range->hmm_pfns[i] &= HMM_PFN_INOUT_FLAGS; in hmm_pfns_fill() 57 range->hmm_pfns[i] |= cpu_flags; in hmm_pfns_fill() 100 struct hmm_range *range = hmm_vma_walk->range; in hmm_pte_need_fault() local 112 pfn_req_flags &= range->pfn_flags_mask; in hmm_pte_need_fault() 113 pfn_req_flags |= range->default_flags; in hmm_pte_need_fault() 135 struct hmm_range *range = hmm_vma_walk->range; in hmm_range_need_fault() local 144 if (!((range->default_flags | range->pfn_flags_mask) & in hmm_range_need_fault() [all …]
|
| H A D | madvise.c | 77 struct madvise_behavior_range range; member 156 struct madvise_behavior_range *range = &madv_behavior->range; in madvise_update_vma() local 159 VMA_ITERATOR(vmi, madv_behavior->mm, range->start); in madvise_update_vma() 167 range->start, range->end, anon_name); in madvise_update_vma() 170 range->start, range->end, &new_flags); in madvise_update_vma() 286 unsigned long start = madv_behavior->range.start; in madvise_willneed() 287 unsigned long end = madv_behavior->range.end; in madvise_willneed() 579 struct madvise_behavior_range *range = &madv_behavior->range; in madvise_cold_page_range() local 586 walk_page_range_vma(vma, range->start, range->end, &cold_walk_ops, in madvise_cold_page_range() 614 struct madvise_behavior_range *range) in madvise_pageout_page_range() argument [all …]
|
| H A D | execmem.c | 28 static void *execmem_vmalloc(struct execmem_range *range, size_t size, in execmem_vmalloc() argument 31 bool kasan = range->flags & EXECMEM_KASAN_SHADOW; in execmem_vmalloc() 33 unsigned int align = range->alignment; in execmem_vmalloc() 34 unsigned long start = range->start; in execmem_vmalloc() 35 unsigned long end = range->end; in execmem_vmalloc() 44 if (!p && range->fallback_start) { in execmem_vmalloc() 45 start = range->fallback_start; in execmem_vmalloc() 46 end = range->fallback_end; in execmem_vmalloc() 67 struct execmem_range *range = &execmem_info->ranges[EXECMEM_MODULE_DATA]; in execmem_vmap() local 70 area = __get_vm_area_node(size, range->alignment, PAGE_SHIFT, VM_ALLOC, in execmem_vmap() [all …]
|
| /linux/include/linux/ |
| H A D | range.h | 6 struct range { struct 11 static inline u64 range_len(const struct range *range) in range_len() argument 13 return range->end - range->start + 1; in range_len() 17 static inline bool range_contains(const struct range *r1, in range_contains() 18 const struct range *r2) in range_contains() 24 static inline bool range_overlaps(const struct range *r1, in range_overlaps() 25 const struct range *r2) in range_overlaps() 30 int add_range(struct range *range, int az, int nr_range, 34 int add_range_with_merge(struct range *range, int az, int nr_range, 37 void subtract_range(struct range *range, int az, u64 start, u64 end); [all …]
|
| /linux/arch/mips/loongson64/ |
| H A D | init.c | 156 struct logic_pio_hwaddr *range; in add_legacy_isa_io() local 159 range = kzalloc(sizeof(*range), GFP_ATOMIC); in add_legacy_isa_io() 160 if (!range) in add_legacy_isa_io() 163 range->fwnode = fwnode; in add_legacy_isa_io() 164 range->size = size = round_up(size, PAGE_SIZE); in add_legacy_isa_io() 165 range->hw_start = hw_start; in add_legacy_isa_io() 166 range->flags = LOGIC_PIO_CPU_MMIO; in add_legacy_isa_io() 168 ret = logic_pio_register_range(range); in add_legacy_isa_io() 170 kfree(range); in add_legacy_isa_io() 175 if (range->io_start != 0) { in add_legacy_isa_io() [all …]
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_svm.c | 22 static bool xe_svm_range_in_vram(struct xe_svm_range *range) in xe_svm_range_in_vram() argument 31 .__flags = READ_ONCE(range->base.pages.flags.__flags), in xe_svm_range_in_vram() 37 static bool xe_svm_range_has_vram_binding(struct xe_svm_range *range) in xe_svm_range_has_vram_binding() argument 40 return xe_svm_range_in_vram(range) && range->tile_present; in xe_svm_range_has_vram_binding() 65 void xe_svm_range_debug(struct xe_svm_range *range, const char *operation) in xe_svm_range_debug() argument 67 range_debug(range, operation); in xe_svm_range_debug() 73 struct xe_svm_range *range; in xe_svm_range_alloc() local 75 range = kzalloc(sizeof(*range), GFP_KERNEL); in xe_svm_range_alloc() 76 if (!range) in xe_svm_range_alloc() 79 INIT_LIST_HEAD(&range->garbage_collector_link); in xe_svm_range_alloc() [all …]
|
| /linux/drivers/pci/hotplug/ |
| H A D | ibmphp_res.c | 368 static int add_bus_range(int type, struct range_node *range, struct bus_node *bus_cur) in add_bus_range() argument 392 if (range->start < range_cur->start) in add_bus_range() 402 bus_cur->rangeMem = range; in add_bus_range() 405 bus_cur->rangePFMem = range; in add_bus_range() 408 bus_cur->rangeIO = range; in add_bus_range() 411 range->next = range_cur; in add_bus_range() 412 range->rangeno = 1; in add_bus_range() 416 range->next = NULL; in add_bus_range() 417 range_prev->next = range; in add_bus_range() 418 range->rangeno = range_prev->rangeno + 1; in add_bus_range() [all …]
|
| /linux/drivers/gpu/drm/sprd/ |
| H A D | megacores_pll.c | 221 u32 range[2], constant; in dphy_timing_config() local 236 range[L] = 50 * scale; in dphy_timing_config() 237 range[H] = INFINITY; in dphy_timing_config() 238 val[CLK] = DIV_ROUND_UP(range[L] * (factor << 1), t_byteck) - 2; in dphy_timing_config() 243 range[L] = 38 * scale; in dphy_timing_config() 244 range[H] = 95 * scale; in dphy_timing_config() 245 tmp = AVERAGE(range[L], range[H]); in dphy_timing_config() 246 val[CLK] = DIV_ROUND_UP(AVERAGE(range[L], range[H]), t_half_byteck) - 1; in dphy_timing_config() 247 range[L] = 40 * scale + 4 * t_ui; in dphy_timing_config() 248 range[H] = 85 * scale + 6 * t_ui; in dphy_timing_config() [all …]
|
| /linux/mm/damon/ |
| H A D | sysfs-common.c | 22 struct damon_sysfs_ul_range *range = kmalloc(sizeof(*range), in damon_sysfs_ul_range_alloc() local 25 if (!range) in damon_sysfs_ul_range_alloc() 27 range->kobj = (struct kobject){}; in damon_sysfs_ul_range_alloc() 28 range->min = min; in damon_sysfs_ul_range_alloc() 29 range->max = max; in damon_sysfs_ul_range_alloc() 31 return range; in damon_sysfs_ul_range_alloc() 37 struct damon_sysfs_ul_range *range = container_of(kobj, in min_show() local 40 return sysfs_emit(buf, "%lu\n", range->min); in min_show() 46 struct damon_sysfs_ul_range *range = container_of(kobj, in min_store() local 55 range->min = min; in min_store() [all …]
|
| /linux/net/netfilter/ |
| H A D | nf_nat_core.c | 392 const struct nf_nat_range2 *range) in nf_nat_used_tuple_harder() 395 return ntohl(t->src.u3.ip) >= ntohl(range->min_addr.ip) && in nf_nat_used_tuple_harder() 396 ntohl(t->src.u3.ip) <= ntohl(range->max_addr.ip); in nf_nat_used_tuple_harder() 398 return ipv6_addr_cmp(&t->src.u3.in6, &range->min_addr.in6) >= 0 && in nf_nat_used_tuple_harder() 399 ipv6_addr_cmp(&t->src.u3.in6, &range->max_addr.in6) <= 0; in nf_nat_used_tuple_harder() 433 * that meet the constraints of range. in l4proto_in_range() 436 const struct nf_nat_range2 *range) in l4proto_in_range() 439 * range specified, otherwise let this drag us onto a new src IP. in l4proto_in_range() 441 if (range->flags & NF_NAT_RANGE_MAP_IPS && in l4proto_in_range() 442 !nf_nat_inet_in_range(tuple, range)) in l4proto_in_range() 404 nf_nat_inet_in_range(const struct nf_conntrack_tuple * t,const struct nf_nat_range2 * range) nf_nat_inet_in_range() argument 448 nf_in_range(const struct nf_conntrack_tuple * tuple,const struct nf_nat_range2 * range) nf_in_range() argument 482 find_appropriate_src(struct net * net,const struct nf_conntrack_zone * zone,const struct nf_conntrack_tuple * tuple,struct nf_conntrack_tuple * result,const struct nf_nat_range2 * range) find_appropriate_src() argument 512 find_best_ips_proto(const struct nf_conntrack_zone * zone,struct nf_conntrack_tuple * tuple,const struct nf_nat_range2 * range,const struct nf_conn * ct,enum nf_nat_manip_type maniptype) find_best_ips_proto() argument 583 nf_nat_l4proto_unique_tuple(struct nf_conntrack_tuple * tuple,const struct nf_nat_range2 * range,enum nf_nat_manip_type maniptype,const struct nf_conn * ct) nf_nat_l4proto_unique_tuple() argument 709 get_unique_tuple(struct nf_conntrack_tuple * tuple,const struct nf_conntrack_tuple * orig_tuple,const struct nf_nat_range2 * range,struct nf_conn * ct,enum nf_nat_manip_type maniptype) get_unique_tuple() argument 784 nf_nat_setup_info(struct nf_conn * ct,const struct nf_nat_range2 * range,enum nf_nat_manip_type maniptype) nf_nat_setup_info() argument 862 struct nf_nat_range2 range = { __nf_nat_alloc_null_binding() local 1048 nf_nat_l4proto_nlattr_to_range(struct nlattr * tb[],struct nf_nat_range2 * range) nf_nat_l4proto_nlattr_to_range() argument 1064 nfnetlink_parse_nat_proto(struct nlattr * attr,const struct nf_conn * ct,struct nf_nat_range2 * range) nfnetlink_parse_nat_proto() argument 1086 nf_nat_ipv4_nlattr_to_range(struct nlattr * tb[],struct nf_nat_range2 * range) nf_nat_ipv4_nlattr_to_range() argument 1100 nf_nat_ipv6_nlattr_to_range(struct nlattr * tb[],struct nf_nat_range2 * range) nf_nat_ipv6_nlattr_to_range() argument 1119 nfnetlink_parse_nat(const struct nlattr * nat,const struct nf_conn * ct,struct nf_nat_range2 * range) nfnetlink_parse_nat() argument 1158 struct nf_nat_range2 range; nfnetlink_parse_nat_setup() local [all...] |
| H A D | xt_nat.c | 55 struct nf_nat_range2 range; in xt_snat_target_v0() local 64 xt_nat_convert_range(&range, &mr->range[0]); in xt_snat_target_v0() 65 return nf_nat_setup_info(ct, &range, NF_NAT_MANIP_SRC); in xt_snat_target_v0() 72 struct nf_nat_range2 range; in xt_dnat_target_v0() local 80 xt_nat_convert_range(&range, &mr->range[0]); in xt_dnat_target_v0() 81 return nf_nat_setup_info(ct, &range, NF_NAT_MANIP_DST); in xt_dnat_target_v0() 88 struct nf_nat_range2 range; in xt_snat_target_v1() local 97 memcpy(&range, range_v1, sizeof(*range_v1)); in xt_snat_target_v1() 98 memset(&range.base_proto, 0, sizeof(range.base_proto)); in xt_snat_target_v1() 100 return nf_nat_setup_info(ct, &range, NF_NAT_MANIP_SRC); in xt_snat_target_v1() [all …]
|
| H A D | nf_nat_bpf.c | 37 struct nf_nat_range2 range; in bpf_ct_set_nat_info() local 42 memset(&range, 0, sizeof(struct nf_nat_range2)); in bpf_ct_set_nat_info() 43 range.flags = NF_NAT_RANGE_MAP_IPS; in bpf_ct_set_nat_info() 44 range.min_addr = *addr; in bpf_ct_set_nat_info() 45 range.max_addr = range.min_addr; in bpf_ct_set_nat_info() 47 range.flags |= NF_NAT_RANGE_PROTO_SPECIFIED; in bpf_ct_set_nat_info() 48 range.min_proto.all = cpu_to_be16(port); in bpf_ct_set_nat_info() 49 range.max_proto.all = range.min_proto.all; in bpf_ct_set_nat_info() 52 return nf_nat_setup_info(ct, &range, manip) == NF_DROP ? -ENOMEM : 0; in bpf_ct_set_nat_info()
|
| H A D | xt_NETMAP.c | 21 const struct nf_nat_range2 *range = par->targinfo; in netmap_tg6() local 29 for (i = 0; i < ARRAY_SIZE(range->min_addr.ip6); i++) in netmap_tg6() 30 netmask.ip6[i] = ~(range->min_addr.ip6[i] ^ in netmap_tg6() 31 range->max_addr.ip6[i]); in netmap_tg6() 41 new_addr.ip6[i] |= range->min_addr.ip6[i] & in netmap_tg6() 45 newrange.flags = range->flags | NF_NAT_RANGE_MAP_IPS; in netmap_tg6() 48 newrange.min_proto = range->min_proto; in netmap_tg6() 49 newrange.max_proto = range->max_proto; in netmap_tg6() 56 const struct nf_nat_range2 *range = par->targinfo; in netmap_tg6_checkentry() local 58 if (!(range->flags & NF_NAT_RANGE_MAP_IPS)) in netmap_tg6_checkentry() [all …]
|
| /linux/tools/testing/selftests/net/ |
| H A D | ip_local_port_range.c | 28 static void unpack_port_range(__u32 range, __u16 *lo, __u16 *hi) in unpack_port_range() argument 30 *lo = range & 0xffff; in unpack_port_range() 31 *hi = range >> 16; in unpack_port_range() 104 static int get_ip_local_port_range(int fd, __u32 *range) in get_ip_local_port_range() argument 115 *range = val; in get_ip_local_port_range() 238 __u32 range; in TEST_F() local 243 range = pack_port_range(t->range_lo, t->range_hi); in TEST_F() 244 err = setsockopt(fd, SOL_IP, IP_LOCAL_PORT_RANGE, &range, sizeof(range)); in TEST_F() 281 __u32 range; in TEST_F() local 289 range = pack_port_range(t->range_lo, t->range_hi); in TEST_F() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlxsw/ |
| H A D | spectrum_port_range.c | 13 struct mlxsw_sp_port_range range; member 36 mlxsw_reg_pprr_src_set(pprr_pl, prr->range.source); in mlxsw_sp_port_range_reg_configure() 37 mlxsw_reg_pprr_dst_set(pprr_pl, !prr->range.source); in mlxsw_sp_port_range_reg_configure() 40 mlxsw_reg_pprr_port_range_min_set(pprr_pl, prr->range.min); in mlxsw_sp_port_range_reg_configure() 41 mlxsw_reg_pprr_port_range_max_set(pprr_pl, prr->range.max); in mlxsw_sp_port_range_reg_configure() 48 const struct mlxsw_sp_port_range *range, in mlxsw_sp_port_range_reg_create() argument 59 prr->range = *range; in mlxsw_sp_port_range_reg_create() 99 const struct mlxsw_sp_port_range *range) in mlxsw_sp_port_range_reg_find() argument 106 if (prr->range.min == range->min && in mlxsw_sp_port_range_reg_find() 107 prr->range.max == range->max && in mlxsw_sp_port_range_reg_find() [all …]
|
| /linux/arch/x86/kernel/cpu/mtrr/ |
| H A D | cleanup.c | 40 static struct range __initdata range[RANGE_NUM]; variable 49 x86_get_mtrr_mem_range(struct range *range, int nr_range, in x86_get_mtrr_mem_range() argument 63 nr_range = add_range_with_merge(range, RANGE_NUM, nr_range, in x86_get_mtrr_mem_range() 70 range[i].start, range[i].end); in x86_get_mtrr_mem_range() 92 subtract_range(range, RANGE_NUM, base, base + size); in x86_get_mtrr_mem_range() 95 subtract_range(range, RANGE_NUM, extra_remove_base, in x86_get_mtrr_mem_range() 100 if (!range[i].end) in x86_get_mtrr_mem_range() 104 range[i].start, range[i].end); in x86_get_mtrr_mem_range() 108 nr_range = clean_sort_range(range, RANGE_NUM); in x86_get_mtrr_mem_range() 113 range[i].start, range[i].end); in x86_get_mtrr_mem_range() [all …]
|
| /linux/arch/s390/boot/ |
| H A D | physmem_info.c | 44 struct physmem_range *range; in add_physmem_online_range() local 47 range = __get_physmem_range_ptr(physmem_info.range_count - 1); in add_physmem_online_range() 48 if (range->end == start) { in add_physmem_online_range() 49 range->end = end; in add_physmem_online_range() 54 range = __get_physmem_range_ptr(physmem_info.range_count); in add_physmem_online_range() 55 range->start = start; in add_physmem_online_range() 56 range->end = end; in add_physmem_online_range() 144 unsigned long range = 1 << (MAX_PHYSMEM_BITS - 20); /* in 1MB blocks */ in search_mem_end() local 148 while (range > 1) { in search_mem_end() 149 range >>= 1; in search_mem_end() [all …]
|
| /linux/drivers/pinctrl/ |
| H A D | core.c | 285 static inline int gpio_to_pin(struct pinctrl_gpio_range *range, in gpio_to_pin() argument 288 unsigned int pin = gc->base + offset - range->base; in gpio_to_pin() 289 if (range->pins) in gpio_to_pin() 290 return range->pins[pin]; in gpio_to_pin() 292 return range->pin_base + pin; in gpio_to_pin() 308 struct pinctrl_gpio_range *range; in pinctrl_match_gpio_range() local 312 list_for_each_entry(range, &pctldev->gpio_ranges, node) { in pinctrl_match_gpio_range() 314 if ((gc->base + offset) >= range->base && in pinctrl_match_gpio_range() 315 (gc->base + offset) < range->base + range->npins) { in pinctrl_match_gpio_range() 317 return range; in pinctrl_match_gpio_range() [all …]
|
| /linux/lib/ |
| H A D | codetag.c | 31 struct codetag_range range; member 66 return cmod->range.start < cmod->range.stop ? cmod->range.start : NULL; in get_first_module_ct() 75 return res < iter->cmod->range.stop ? res : NULL; in get_next_module_ct() 132 const struct codetag_range *range) in range_size() argument 134 return ((char *)range->stop - (char *)range->start) / in range_size() 178 struct codetag_range range; in codetag_module_init() local 183 range = get_section_range(mod, cttype->desc.section); in codetag_module_init() 184 if (!range.start || !range.stop) { in codetag_module_init() 191 if (range.start == range.stop) in codetag_module_init() 194 BUG_ON(range.start > range.stop); in codetag_module_init() [all …]
|