Home
last modified time | relevance | path

Searched refs:range (Results 26 – 50 of 1733) sorted by relevance

12345678910>>...70

/linux/Documentation/translations/zh_CN/mm/
H A Dhmm.rst129 int hmm_range_fault(struct hmm_range *range);
142 struct hmm_range range;
145 range.notifier = &interval_sub;
146 range.start = ...;
147 range.end = ...;
148 range.hmm_pfns = ...;
154 range.notifier_seq = mmu_interval_read_begin(&interval_sub);
156 ret = hmm_range_fault(&range);
166 if (mmu_interval_read_retry(&ni, range.notifier_seq) {
189 range->default_flags = HMM_PFN_REQ_FAULT;
[all …]
/linux/drivers/regulator/
H A Dqcom_spmi-regulator.c361 /* VSET value to decide the range of ULT SMPS */
373 * to pick which range should be used in the case of
376 * range
377 * @range_sel: Voltage range register value corresponding to this range
386 * specify that the voltage range has meaning, but is not preferred.
400 * so that range[i].set_point_max_uV < range[i+1].set_point_min_uV.
403 struct spmi_voltage_range *range; member
481 .range
669 const struct spmi_voltage_range *range; spmi_regulator_select_voltage() local
726 const struct spmi_voltage_range *range, *end; spmi_sw_selector_to_hw() local
752 spmi_hw_selector_to_sw(struct spmi_regulator * vreg,u8 hw_sel,const struct spmi_voltage_range * range) spmi_hw_selector_to_sw() argument
789 const struct spmi_voltage_range *range, *end; spmi_regulator_find_range() local
806 const struct spmi_voltage_range *range; spmi_regulator_select_voltage_same_range() local
920 const struct spmi_voltage_range *range; spmi_regulator_common_get_voltage() local
935 const struct spmi_voltage_range *range; spmi_regulator_ftsmps426_get_voltage() local
1009 const struct spmi_voltage_range *range; spmi_regulator_ult_lo_smps_get_voltage() local
1680 struct spmi_voltage_range *range = points->range; spmi_calculate_num_voltages() local
1754 const struct spmi_voltage_range *range; spmi_regulator_init_slew_rate() local
1799 const struct spmi_voltage_range *range = &vreg->set_points->range[0]; spmi_regulator_init_slew_rate_ftsmps426() local
2504 const struct spmi_voltage_range *range; qcom_spmi_regulator_probe() local
[all...]
/linux/net/netfilter/
H A Dnft_nat.c33 static void nft_nat_setup_addr(struct nf_nat_range2 *range, in nft_nat_setup_addr() argument
39 range->min_addr.ip = (__force __be32) in nft_nat_setup_addr()
41 range->max_addr.ip = (__force __be32) in nft_nat_setup_addr()
45 memcpy(range->min_addr.ip6, &regs->data[priv->sreg_addr_min], in nft_nat_setup_addr()
46 sizeof(range->min_addr.ip6)); in nft_nat_setup_addr()
47 memcpy(range->max_addr.ip6, &regs->data[priv->sreg_addr_max], in nft_nat_setup_addr()
48 sizeof(range->max_addr.ip6)); in nft_nat_setup_addr()
53 static void nft_nat_setup_proto(struct nf_nat_range2 *range, in nft_nat_setup_proto() argument
57 range->min_proto.all = (__force __be16) in nft_nat_setup_proto()
59 range->max_proto.all = (__force __be16) in nft_nat_setup_proto()
[all …]
H A Dnf_nat_helper.c182 struct nf_nat_range2 range; in nf_nat_follow_master() local
188 range.flags = NF_NAT_RANGE_MAP_IPS; in nf_nat_follow_master()
189 range.min_addr = range.max_addr in nf_nat_follow_master()
191 nf_nat_setup_info(ct, &range, NF_NAT_MANIP_SRC); in nf_nat_follow_master()
194 range.flags = (NF_NAT_RANGE_MAP_IPS | NF_NAT_RANGE_PROTO_SPECIFIED); in nf_nat_follow_master()
195 range.min_proto = range.max_proto = exp->saved_proto; in nf_nat_follow_master()
196 range.min_addr = range.max_addr in nf_nat_follow_master()
198 nf_nat_setup_info(ct, &range, NF_NAT_MANIP_DST); in nf_nat_follow_master()
205 int range, attempts_left; in nf_nat_exp_find_port() local
208 range = USHRT_MAX - port; in nf_nat_exp_find_port()
[all …]
H A Dxt_MASQUERADE.c23 if (mr->range[0].flags & NF_NAT_RANGE_MAP_IPS) { in masquerade_tg_check()
37 struct nf_nat_range2 range; in masquerade_tg() local
41 range.flags = mr->range[0].flags; in masquerade_tg()
42 range.min_proto = mr->range[0].min; in masquerade_tg()
43 range.max_proto = mr->range[0].max; in masquerade_tg()
45 return nf_nat_masquerade_ipv4(skb, xt_hooknum(par), &range, in masquerade_tg()
63 const struct nf_nat_range2 *range = par->targinfo; in masquerade_tg6_checkentry() local
65 if (range->flags & NF_NAT_RANGE_MAP_IPS) in masquerade_tg6_checkentry()
H A Dxt_REDIRECT.c36 const struct nf_nat_range2 *range = par->targinfo; in redirect_tg6_checkentry() local
38 if (range->flags & NF_NAT_RANGE_MAP_IPS) in redirect_tg6_checkentry()
53 if (mr->range[0].flags & NF_NAT_RANGE_MAP_IPS) { in redirect_tg4_check()
68 struct nf_nat_range2 range = { in redirect_tg4() local
69 .flags = mr->range[0].flags, in redirect_tg4()
70 .min_proto = mr->range[0].min, in redirect_tg4()
71 .max_proto = mr->range[0].max, in redirect_tg4()
74 return nf_nat_redirect_ipv4(skb, &range, xt_hooknum(par)); in redirect_tg4()
/linux/tools/testing/selftests/bpf/prog_tests/
H A Dreg_bounds.c218 struct range { struct
222 static void snprintf_range(enum num_t t, struct strbuf *sb, struct range x) in snprintf_range() argument
234 static void print_range(enum num_t t, struct range x, const char *sfx) in print_range()
242 static const struct range unkn[] = {
249 static struct range unkn_subreg(enum num_t t) in unkn_subreg()
260 static struct range range(enum num_t t, u64 a, u64 b) in range() function
263 case U64: return (struct range){ (u64)a, (u64)b }; in range()
264 case U32: return (struct range){ (u32)a, (u32)b }; in range()
265 case S64: return (struct range){ (s6 in range()
[all...]
/linux/drivers/base/
H A Dmap.c23 unsigned long range; member
32 int kobj_map(struct kobj_map *domain, dev_t dev, unsigned long range, in kobj_map() argument
36 unsigned int n = MAJOR(dev + range - 1) - MAJOR(dev) + 1; in kobj_map()
53 p->range = range; in kobj_map()
59 while (*s && (*s)->range < range) in kobj_map()
68 void kobj_unmap(struct kobj_map *domain, dev_t dev, unsigned long range) in kobj_unmap() argument
70 unsigned int n = MAJOR(dev + range - 1) - MAJOR(dev) + 1; in kobj_unmap()
83 if (p->dev == dev && p->range == range) { in kobj_unmap()
108 if (p->dev > dev || p->dev + p->range - 1 < dev) in kobj_lookup()
110 if (p->range - 1 >= best) in kobj_lookup()
[all …]
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_uvmm.c69 u64 range; member
86 u64 range; member
92 u64 addr, u64 range) in nouveau_uvmm_vmm_sparse_ref() argument
96 return nvif_vmm_raw_sparse(vmm, addr, range, true); in nouveau_uvmm_vmm_sparse_ref()
101 u64 addr, u64 range) in nouveau_uvmm_vmm_sparse_unref() argument
105 return nvif_vmm_raw_sparse(vmm, addr, range, false); in nouveau_uvmm_vmm_sparse_unref()
110 u64 addr, u64 range) in nouveau_uvmm_vmm_get() argument
114 return nvif_vmm_raw_get(vmm, addr, range, PAGE_SHIFT); in nouveau_uvmm_vmm_get()
119 u64 addr, u64 range) in nouveau_uvmm_vmm_put() argument
123 return nvif_vmm_raw_put(vmm, addr, range, PAGE_SHIFT); in nouveau_uvmm_vmm_put()
[all …]
/linux/arch/x86/kvm/
H A Dkvm_onhyperv.c21 struct kvm_hv_tlb_range *range = data; in kvm_fill_hv_flush_list_func() local
23 return hyperv_fill_flush_guest_mapping_list(flush, range->start_gfn, in kvm_fill_hv_flush_list_func()
24 range->pages); in kvm_fill_hv_flush_list_func()
28 struct kvm_hv_tlb_range *range) in hv_remote_flush_root_tdp() argument
30 if (range) in hv_remote_flush_root_tdp()
32 kvm_fill_hv_flush_list_func, (void *)range); in hv_remote_flush_root_tdp()
38 struct kvm_hv_tlb_range *range) in __hv_flush_remote_tlbs_range() argument
71 ret = hv_remote_flush_root_tdp(root, range); in __hv_flush_remote_tlbs_range()
88 ret = hv_remote_flush_root_tdp(kvm_arch->hv_root_tdp, range); in __hv_flush_remote_tlbs_range()
97 struct kvm_hv_tlb_range range = { in hv_flush_remote_tlbs_range() local
[all …]
/linux/net/phonet/
H A Dsysctl.c30 static void set_local_port_range(int range[2]) in set_local_port_range()
33 local_port_range[0] = range[0]; in set_local_port_range()
34 local_port_range[1] = range[1]; in set_local_port_range()
55 int range[2] = {local_port_range[0], local_port_range[1]}; in proc_local_port_range() local
57 .data = &range, in proc_local_port_range()
58 .maxlen = sizeof(range), in proc_local_port_range()
67 if (range[1] < range[0]) in proc_local_port_range()
70 set_local_port_range(range); in proc_local_port_range()
/linux/mm/
H A Dmmu_notifier.c71 * - some range on the mm_struct is being invalidated
77 * - some range on the mm_struct is being invalidated
97 const struct mmu_notifier_range *range, in mn_itree_inv_start_range()
105 node = interval_tree_iter_first(&subscriptions->itree, range->start, in mn_itree_inv_start_range()
106 range->end - 1); in mn_itree_inv_start_range()
120 const struct mmu_notifier_range *range) in mn_itree_inv_next()
125 range->start, range->end - 1); in mn_itree_inv_next()
170 * range
174 * collision-retry scheme similar to seqcount for the VA range unde
95 mn_itree_inv_start_range(struct mmu_notifier_subscriptions * subscriptions,const struct mmu_notifier_range * range,unsigned long * seq) mn_itree_inv_start_range() argument
118 mn_itree_inv_next(struct mmu_interval_notifier * interval_sub,const struct mmu_notifier_range * range) mn_itree_inv_next() argument
264 struct mmu_notifier_range range = { mn_itree_release() local
428 mn_itree_invalidate(struct mmu_notifier_subscriptions * subscriptions,const struct mmu_notifier_range * range) mn_itree_invalidate() argument
460 mn_hlist_invalidate_range_start(struct mmu_notifier_subscriptions * subscriptions,struct mmu_notifier_range * range) mn_hlist_invalidate_range_start() argument
519 __mmu_notifier_invalidate_range_start(struct mmu_notifier_range * range) __mmu_notifier_invalidate_range_start() argument
537 mn_hlist_invalidate_end(struct mmu_notifier_subscriptions * subscriptions,struct mmu_notifier_range * range) mn_hlist_invalidate_end() argument
557 __mmu_notifier_invalidate_range_end(struct mmu_notifier_range * range) __mmu_notifier_invalidate_range_end() argument
[all...]
H A Dexecmem.c17 static void *__execmem_alloc(struct execmem_range *range, size_t size) in __execmem_alloc() argument
19 bool kasan = range->flags & EXECMEM_KASAN_SHADOW; in __execmem_alloc()
22 unsigned long start = range->start; in __execmem_alloc()
23 unsigned long end = range->end; in __execmem_alloc()
24 unsigned int align = range->alignment; in __execmem_alloc()
25 pgprot_t pgprot = range->pgprot; in __execmem_alloc()
34 if (!p && range->fallback_start) { in __execmem_alloc()
35 start = range->fallback_start; in __execmem_alloc()
36 end = range->fallback_end; in __execmem_alloc()
57 struct execmem_range *range = &execmem_info->ranges[type]; in execmem_alloc() local
[all …]
/linux/arch/x86/pci/
H A Damd_bus.c81 struct range range[RANGE_NUM]; in early_root_info_init() local
157 memset(range, 0, sizeof(range)); in early_root_info_init()
158 add_range(range, RANGE_NUM, 0, 0, 0xffff + 1); in early_root_info_init()
182 subtract_range(range, RANGE_NUM, start, end + 1); in early_root_info_init()
189 if (!range[i].end) in early_root_info_init()
192 update_res(info, range[i].start, range[i].end - 1, in early_root_info_init()
197 memset(range, 0, sizeof(range)); in early_root_info_init()
201 add_range(range, RANGE_NUM, 0, 0, end); in early_root_info_init()
209 subtract_range(range, RANGE_NUM, 0, end); in early_root_info_init()
218 subtract_range(range, RANGE_NUM, fam10h_mmconf_start, in early_root_info_init()
[all …]
/linux/drivers/virt/acrn/
H A Dioreq.c118 struct acrn_ioreq_range *range; in acrn_ioreq_range_add() local
126 range = kzalloc(sizeof(*range), GFP_KERNEL); in acrn_ioreq_range_add()
127 if (!range) in acrn_ioreq_range_add()
130 range->type = type; in acrn_ioreq_range_add()
131 range->start = start; in acrn_ioreq_range_add()
132 range->end = end; in acrn_ioreq_range_add()
135 list_add(&range->list, &client->range_list); in acrn_ioreq_range_add()
151 struct acrn_ioreq_range *range; in acrn_ioreq_range_del() local
154 list_for_each_entry(range, &client->range_list, list) { in acrn_ioreq_range_del()
155 if (type == range->type && in acrn_ioreq_range_del()
[all …]
/linux/drivers/nvdimm/
H A Dbadrange.c214 struct badblocks *bb, const struct range *range) in badblocks_populate() argument
225 if (bre_end < range->start) in badblocks_populate()
227 if (bre->start > range->end) in badblocks_populate()
230 if (bre->start >= range->start) { in badblocks_populate()
234 if (bre_end <= range->end) in badblocks_populate()
237 len = range->start + range_len(range) in badblocks_populate()
239 __add_badblock_range(bb, start - range->start, len); in badblocks_populate()
246 if (bre->start < range->start) { in badblocks_populate()
249 if (bre_end < range->end) in badblocks_populate()
250 len = bre->start + bre->length - range->start; in badblocks_populate()
[all …]
/linux/Documentation/userspace-api/media/v4l/
H A Dcolorspaces-details.rst17 range. The chromaticities of the primary colors and the white reference
77 Y' is clamped to the range [0…1] and Cb and Cr are clamped to the range
82 The default quantization is limited range, but full range is possible
94 quantization is limited range. The chromaticities of the primary colors
121 Transfer function. Normally L is in the range [0…1], but for the
122 extended gamut xvYCC encoding values outside that range are allowed.
153 Y' is clamped to the range [0…1] and Cb and Cr are clamped to the range
156 The default quantization is limited range, but full range is possible
168 that are outside the range [0…1]. The resulting Y', Cb and Cr values are
169 scaled and offset according to the limited range formula:
[all …]
/linux/lib/
H A Dtest_hmm.c207 static int dmirror_do_fault(struct dmirror *dmirror, struct hmm_range *range) in dmirror_do_fault() argument
209 unsigned long *pfns = range->hmm_pfns; in dmirror_do_fault()
212 for (pfn = (range->start >> PAGE_SHIFT); in dmirror_do_fault()
213 pfn < (range->end >> PAGE_SHIFT); in dmirror_do_fault()
231 else if (WARN_ON(range->default_flags & HMM_PFN_WRITE)) in dmirror_do_fault()
258 const struct mmu_notifier_range *range, in dmirror_interval_invalidate() argument
267 if (range->event == MMU_NOTIFY_MIGRATE && in dmirror_interval_invalidate()
268 range->owner == dmirror->mdevice) in dmirror_interval_invalidate()
271 if (mmu_notifier_range_blockable(range)) in dmirror_interval_invalidate()
277 dmirror_do_update(dmirror, range in dmirror_interval_invalidate()
288 dmirror_range_fault(struct dmirror * dmirror,struct hmm_range * range) dmirror_range_fault() argument
333 struct hmm_range range = { dmirror_fault() local
1039 dmirror_mkentry(struct dmirror * dmirror,struct hmm_range * range,unsigned char * perm,unsigned long entry) dmirror_mkentry() argument
1081 dmirror_snapshot_invalidate(struct mmu_interval_notifier * mni,const struct mmu_notifier_range * range,unsigned long cur_seq) dmirror_snapshot_invalidate() argument
1108 dmirror_range_snapshot(struct dmirror * dmirror,struct hmm_range * range,unsigned char * perm) dmirror_range_snapshot() argument
1175 struct hmm_range range = { dmirror_snapshot() local
[all...]
H A Dnlattr.c120 struct netlink_range_validation *range) in nla_get_range_unsigned() argument
125 range->min = 0; in nla_get_range_unsigned()
129 range->max = U8_MAX; in nla_get_range_unsigned()
134 range->max = U16_MAX; in nla_get_range_unsigned()
138 range->max = U32_MAX; in nla_get_range_unsigned()
143 range->max = U64_MAX; in nla_get_range_unsigned()
153 range->min = pt->min; in nla_get_range_unsigned()
154 range->max = pt->max; in nla_get_range_unsigned()
157 *range = *pt->range; in nla_get_range_unsigned()
160 range->min = pt->min; in nla_get_range_unsigned()
[all …]
/linux/drivers/gpu/drm/logicvc/
H A Dlogicvc_of.c42 .range = {
50 .range = {
57 .range = { 8, 24 },
79 .range = { 8, 24 },
84 .range = {
92 .range = {
158 if (property->range[0] || property->range[1]) in logicvc_of_property_parse_u32()
159 if (value < property->range[0] || value > property->range[1]) in logicvc_of_property_parse_u32()
/linux/include/linux/comedi/
H A Dcomedidev.h636 struct comedi_krange range[] __counted_by(length);
655 unsigned int range) in comedi_range_is_bipolar() argument
657 return s->range_table->range[range].min < 0; in comedi_range_is_bipolar()
676 unsigned int range) in comedi_range_is_unipolar() argument
678 return s->range_table->range[range].min >= 0; in comedi_range_is_unipolar()
697 unsigned int range) in comedi_range_is_external() argument
699 return !!(s->range_table->range[range].flags & RF_EXTERNAL); in comedi_range_is_external()
720 unsigned int range) in comedi_chan_range_is_bipolar() argument
722 return s->range_table_list[chan]->range[range].min < 0; in comedi_chan_range_is_bipolar()
743 unsigned int range) in comedi_chan_range_is_unipolar() argument
[all …]
/linux/include/drm/
H A Ddrm_gpuvm.h102 u64 range; member
157 u64 addr, u64 range);
159 u64 addr, u64 range);
163 static inline void drm_gpuva_init(struct drm_gpuva *va, u64 addr, u64 range, in drm_gpuva_init() argument
167 va->va.range = range; in drm_gpuva_init()
339 u64 start_offset, u64 range,
362 bool drm_gpuvm_range_valid(struct drm_gpuvm *gpuvm, u64 addr, u64 range);
363 bool drm_gpuvm_interval_empty(struct drm_gpuvm *gpuvm, u64 addr, u64 range);
562 u64 addr, u64 range,
572 u64 addr, u64 range);
[all …]
/linux/drivers/comedi/drivers/
H A Djr3_pci.c94 char _reserved[offsetof(struct comedi_lrange, range[1])];
112 union jr3_pci_single_range range[9]; member
508 union jr3_pci_single_range *r = spriv->range; in jr3_pci_poll_subdevice()
511 r[0].l.range[0].min = -get_s16(&fs->fx) * 1000; in jr3_pci_poll_subdevice()
512 r[0].l.range[0].max = get_s16(&fs->fx) * 1000; in jr3_pci_poll_subdevice()
513 r[1].l.range[0].min = -get_s16(&fs->fy) * 1000; in jr3_pci_poll_subdevice()
514 r[1].l.range[0].max = get_s16(&fs->fy) * 1000; in jr3_pci_poll_subdevice()
515 r[2].l.range[0].min = -get_s16(&fs->fz) * 1000; in jr3_pci_poll_subdevice()
516 r[2].l.range[0].max = get_s16(&fs->fz) * 1000; in jr3_pci_poll_subdevice()
517 r[3].l.range[0].min = -get_s16(&fs->mx) * 100; in jr3_pci_poll_subdevice()
[all …]
/linux/drivers/infiniband/core/
H A Dumem_odp.c355 struct hmm_range range = {}; in ib_umem_odp_map_dma_and_lock() local
378 range.notifier = &umem_odp->notifier; in ib_umem_odp_map_dma_and_lock()
379 range.start = ALIGN_DOWN(user_virt, 1UL << page_shift); in ib_umem_odp_map_dma_and_lock()
380 range.end = ALIGN(user_virt + bcnt, 1UL << page_shift); in ib_umem_odp_map_dma_and_lock()
381 pfn_start_idx = (range.start - ib_umem_start(umem_odp)) >> PAGE_SHIFT; in ib_umem_odp_map_dma_and_lock()
382 num_pfns = (range.end - range.start) >> PAGE_SHIFT; in ib_umem_odp_map_dma_and_lock()
384 range.default_flags = HMM_PFN_REQ_FAULT; in ib_umem_odp_map_dma_and_lock()
387 range.default_flags |= HMM_PFN_REQ_WRITE; in ib_umem_odp_map_dma_and_lock()
390 range.hmm_pfns = &(umem_odp->pfn_list[pfn_start_idx]); in ib_umem_odp_map_dma_and_lock()
394 current_seq = range.notifier_seq = in ib_umem_odp_map_dma_and_lock()
[all …]
/linux/drivers/gpu/drm/
H A Ddrm_gpuvm.c871 #define GPUVA_LAST(node) ((node)->va.addr + (node)->va.range - 1)
885 drm_gpuvm_check_overflow(u64 addr, u64 range) in drm_gpuvm_check_overflow() argument
889 return check_add_overflow(addr, range, &end); in drm_gpuvm_check_overflow()
893 drm_gpuvm_warn_check_overflow(struct drm_gpuvm *gpuvm, u64 addr, u64 range) in drm_gpuvm_warn_check_overflow() argument
895 return drm_WARN(gpuvm->drm, drm_gpuvm_check_overflow(addr, range), in drm_gpuvm_warn_check_overflow()
900 drm_gpuvm_in_mm_range(struct drm_gpuvm *gpuvm, u64 addr, u64 range) in drm_gpuvm_in_mm_range() argument
902 u64 end = addr + range; in drm_gpuvm_in_mm_range()
910 drm_gpuvm_in_kernel_node(struct drm_gpuvm *gpuvm, u64 addr, u64 range) in drm_gpuvm_in_kernel_node() argument
912 u64 end = addr + range; in drm_gpuvm_in_kernel_node()
914 u64 krange = gpuvm->kernel_alloc_node.va.range; in drm_gpuvm_in_kernel_node()
[all …]

12345678910>>...70