/linux/Documentation/translations/zh_CN/mm/ |
H A D | hmm.rst | 129 int hmm_range_fault(struct hmm_range *range); 142 struct hmm_range range; 145 range.notifier = &interval_sub; 146 range.start = ...; 147 range.end = ...; 148 range.hmm_pfns = ...; 154 range.notifier_seq = mmu_interval_read_begin(&interval_sub); 156 ret = hmm_range_fault(&range); 166 if (mmu_interval_read_retry(&ni, range.notifier_seq) { 189 range->default_flags = HMM_PFN_REQ_FAULT; [all …]
|
/linux/drivers/regulator/ |
H A D | qcom_spmi-regulator.c | 361 /* VSET value to decide the range of ULT SMPS */ 373 * to pick which range should be used in the case of 376 * range 377 * @range_sel: Voltage range register value corresponding to this range 386 * specify that the voltage range has meaning, but is not preferred. 400 * so that range[i].set_point_max_uV < range[i+1].set_point_min_uV. 403 struct spmi_voltage_range *range; member 481 .range 669 const struct spmi_voltage_range *range; spmi_regulator_select_voltage() local 726 const struct spmi_voltage_range *range, *end; spmi_sw_selector_to_hw() local 752 spmi_hw_selector_to_sw(struct spmi_regulator * vreg,u8 hw_sel,const struct spmi_voltage_range * range) spmi_hw_selector_to_sw() argument 789 const struct spmi_voltage_range *range, *end; spmi_regulator_find_range() local 806 const struct spmi_voltage_range *range; spmi_regulator_select_voltage_same_range() local 920 const struct spmi_voltage_range *range; spmi_regulator_common_get_voltage() local 935 const struct spmi_voltage_range *range; spmi_regulator_ftsmps426_get_voltage() local 1009 const struct spmi_voltage_range *range; spmi_regulator_ult_lo_smps_get_voltage() local 1680 struct spmi_voltage_range *range = points->range; spmi_calculate_num_voltages() local 1754 const struct spmi_voltage_range *range; spmi_regulator_init_slew_rate() local 1799 const struct spmi_voltage_range *range = &vreg->set_points->range[0]; spmi_regulator_init_slew_rate_ftsmps426() local 2504 const struct spmi_voltage_range *range; qcom_spmi_regulator_probe() local [all...] |
/linux/net/netfilter/ |
H A D | nft_nat.c | 33 static void nft_nat_setup_addr(struct nf_nat_range2 *range, in nft_nat_setup_addr() argument 39 range->min_addr.ip = (__force __be32) in nft_nat_setup_addr() 41 range->max_addr.ip = (__force __be32) in nft_nat_setup_addr() 45 memcpy(range->min_addr.ip6, ®s->data[priv->sreg_addr_min], in nft_nat_setup_addr() 46 sizeof(range->min_addr.ip6)); in nft_nat_setup_addr() 47 memcpy(range->max_addr.ip6, ®s->data[priv->sreg_addr_max], in nft_nat_setup_addr() 48 sizeof(range->max_addr.ip6)); in nft_nat_setup_addr() 53 static void nft_nat_setup_proto(struct nf_nat_range2 *range, in nft_nat_setup_proto() argument 57 range->min_proto.all = (__force __be16) in nft_nat_setup_proto() 59 range->max_proto.all = (__force __be16) in nft_nat_setup_proto() [all …]
|
H A D | nf_nat_helper.c | 182 struct nf_nat_range2 range; in nf_nat_follow_master() local 188 range.flags = NF_NAT_RANGE_MAP_IPS; in nf_nat_follow_master() 189 range.min_addr = range.max_addr in nf_nat_follow_master() 191 nf_nat_setup_info(ct, &range, NF_NAT_MANIP_SRC); in nf_nat_follow_master() 194 range.flags = (NF_NAT_RANGE_MAP_IPS | NF_NAT_RANGE_PROTO_SPECIFIED); in nf_nat_follow_master() 195 range.min_proto = range.max_proto = exp->saved_proto; in nf_nat_follow_master() 196 range.min_addr = range.max_addr in nf_nat_follow_master() 198 nf_nat_setup_info(ct, &range, NF_NAT_MANIP_DST); in nf_nat_follow_master() 205 int range, attempts_left; in nf_nat_exp_find_port() local 208 range = USHRT_MAX - port; in nf_nat_exp_find_port() [all …]
|
H A D | xt_MASQUERADE.c | 23 if (mr->range[0].flags & NF_NAT_RANGE_MAP_IPS) { in masquerade_tg_check() 37 struct nf_nat_range2 range; in masquerade_tg() local 41 range.flags = mr->range[0].flags; in masquerade_tg() 42 range.min_proto = mr->range[0].min; in masquerade_tg() 43 range.max_proto = mr->range[0].max; in masquerade_tg() 45 return nf_nat_masquerade_ipv4(skb, xt_hooknum(par), &range, in masquerade_tg() 63 const struct nf_nat_range2 *range = par->targinfo; in masquerade_tg6_checkentry() local 65 if (range->flags & NF_NAT_RANGE_MAP_IPS) in masquerade_tg6_checkentry()
|
H A D | xt_REDIRECT.c | 36 const struct nf_nat_range2 *range = par->targinfo; in redirect_tg6_checkentry() local 38 if (range->flags & NF_NAT_RANGE_MAP_IPS) in redirect_tg6_checkentry() 53 if (mr->range[0].flags & NF_NAT_RANGE_MAP_IPS) { in redirect_tg4_check() 68 struct nf_nat_range2 range = { in redirect_tg4() local 69 .flags = mr->range[0].flags, in redirect_tg4() 70 .min_proto = mr->range[0].min, in redirect_tg4() 71 .max_proto = mr->range[0].max, in redirect_tg4() 74 return nf_nat_redirect_ipv4(skb, &range, xt_hooknum(par)); in redirect_tg4()
|
/linux/tools/testing/selftests/bpf/prog_tests/ |
H A D | reg_bounds.c | 218 struct range { struct 222 static void snprintf_range(enum num_t t, struct strbuf *sb, struct range x) in snprintf_range() argument 234 static void print_range(enum num_t t, struct range x, const char *sfx) in print_range() 242 static const struct range unkn[] = { 249 static struct range unkn_subreg(enum num_t t) in unkn_subreg() 260 static struct range range(enum num_t t, u64 a, u64 b) in range() function 263 case U64: return (struct range){ (u64)a, (u64)b }; in range() 264 case U32: return (struct range){ (u32)a, (u32)b }; in range() 265 case S64: return (struct range){ (s6 in range() [all...] |
/linux/drivers/base/ |
H A D | map.c | 23 unsigned long range; member 32 int kobj_map(struct kobj_map *domain, dev_t dev, unsigned long range, in kobj_map() argument 36 unsigned int n = MAJOR(dev + range - 1) - MAJOR(dev) + 1; in kobj_map() 53 p->range = range; in kobj_map() 59 while (*s && (*s)->range < range) in kobj_map() 68 void kobj_unmap(struct kobj_map *domain, dev_t dev, unsigned long range) in kobj_unmap() argument 70 unsigned int n = MAJOR(dev + range - 1) - MAJOR(dev) + 1; in kobj_unmap() 83 if (p->dev == dev && p->range == range) { in kobj_unmap() 108 if (p->dev > dev || p->dev + p->range - 1 < dev) in kobj_lookup() 110 if (p->range - 1 >= best) in kobj_lookup() [all …]
|
/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_uvmm.c | 69 u64 range; member 86 u64 range; member 92 u64 addr, u64 range) in nouveau_uvmm_vmm_sparse_ref() argument 96 return nvif_vmm_raw_sparse(vmm, addr, range, true); in nouveau_uvmm_vmm_sparse_ref() 101 u64 addr, u64 range) in nouveau_uvmm_vmm_sparse_unref() argument 105 return nvif_vmm_raw_sparse(vmm, addr, range, false); in nouveau_uvmm_vmm_sparse_unref() 110 u64 addr, u64 range) in nouveau_uvmm_vmm_get() argument 114 return nvif_vmm_raw_get(vmm, addr, range, PAGE_SHIFT); in nouveau_uvmm_vmm_get() 119 u64 addr, u64 range) in nouveau_uvmm_vmm_put() argument 123 return nvif_vmm_raw_put(vmm, addr, range, PAGE_SHIFT); in nouveau_uvmm_vmm_put() [all …]
|
/linux/arch/x86/kvm/ |
H A D | kvm_onhyperv.c | 21 struct kvm_hv_tlb_range *range = data; in kvm_fill_hv_flush_list_func() local 23 return hyperv_fill_flush_guest_mapping_list(flush, range->start_gfn, in kvm_fill_hv_flush_list_func() 24 range->pages); in kvm_fill_hv_flush_list_func() 28 struct kvm_hv_tlb_range *range) in hv_remote_flush_root_tdp() argument 30 if (range) in hv_remote_flush_root_tdp() 32 kvm_fill_hv_flush_list_func, (void *)range); in hv_remote_flush_root_tdp() 38 struct kvm_hv_tlb_range *range) in __hv_flush_remote_tlbs_range() argument 71 ret = hv_remote_flush_root_tdp(root, range); in __hv_flush_remote_tlbs_range() 88 ret = hv_remote_flush_root_tdp(kvm_arch->hv_root_tdp, range); in __hv_flush_remote_tlbs_range() 97 struct kvm_hv_tlb_range range = { in hv_flush_remote_tlbs_range() local [all …]
|
/linux/net/phonet/ |
H A D | sysctl.c | 30 static void set_local_port_range(int range[2]) in set_local_port_range() 33 local_port_range[0] = range[0]; in set_local_port_range() 34 local_port_range[1] = range[1]; in set_local_port_range() 55 int range[2] = {local_port_range[0], local_port_range[1]}; in proc_local_port_range() local 57 .data = &range, in proc_local_port_range() 58 .maxlen = sizeof(range), in proc_local_port_range() 67 if (range[1] < range[0]) in proc_local_port_range() 70 set_local_port_range(range); in proc_local_port_range()
|
/linux/mm/ |
H A D | mmu_notifier.c | 71 * - some range on the mm_struct is being invalidated 77 * - some range on the mm_struct is being invalidated 97 const struct mmu_notifier_range *range, in mn_itree_inv_start_range() 105 node = interval_tree_iter_first(&subscriptions->itree, range->start, in mn_itree_inv_start_range() 106 range->end - 1); in mn_itree_inv_start_range() 120 const struct mmu_notifier_range *range) in mn_itree_inv_next() 125 range->start, range->end - 1); in mn_itree_inv_next() 170 * range 174 * collision-retry scheme similar to seqcount for the VA range unde 95 mn_itree_inv_start_range(struct mmu_notifier_subscriptions * subscriptions,const struct mmu_notifier_range * range,unsigned long * seq) mn_itree_inv_start_range() argument 118 mn_itree_inv_next(struct mmu_interval_notifier * interval_sub,const struct mmu_notifier_range * range) mn_itree_inv_next() argument 264 struct mmu_notifier_range range = { mn_itree_release() local 428 mn_itree_invalidate(struct mmu_notifier_subscriptions * subscriptions,const struct mmu_notifier_range * range) mn_itree_invalidate() argument 460 mn_hlist_invalidate_range_start(struct mmu_notifier_subscriptions * subscriptions,struct mmu_notifier_range * range) mn_hlist_invalidate_range_start() argument 519 __mmu_notifier_invalidate_range_start(struct mmu_notifier_range * range) __mmu_notifier_invalidate_range_start() argument 537 mn_hlist_invalidate_end(struct mmu_notifier_subscriptions * subscriptions,struct mmu_notifier_range * range) mn_hlist_invalidate_end() argument 557 __mmu_notifier_invalidate_range_end(struct mmu_notifier_range * range) __mmu_notifier_invalidate_range_end() argument [all...] |
H A D | execmem.c | 17 static void *__execmem_alloc(struct execmem_range *range, size_t size) in __execmem_alloc() argument 19 bool kasan = range->flags & EXECMEM_KASAN_SHADOW; in __execmem_alloc() 22 unsigned long start = range->start; in __execmem_alloc() 23 unsigned long end = range->end; in __execmem_alloc() 24 unsigned int align = range->alignment; in __execmem_alloc() 25 pgprot_t pgprot = range->pgprot; in __execmem_alloc() 34 if (!p && range->fallback_start) { in __execmem_alloc() 35 start = range->fallback_start; in __execmem_alloc() 36 end = range->fallback_end; in __execmem_alloc() 57 struct execmem_range *range = &execmem_info->ranges[type]; in execmem_alloc() local [all …]
|
/linux/arch/x86/pci/ |
H A D | amd_bus.c | 81 struct range range[RANGE_NUM]; in early_root_info_init() local 157 memset(range, 0, sizeof(range)); in early_root_info_init() 158 add_range(range, RANGE_NUM, 0, 0, 0xffff + 1); in early_root_info_init() 182 subtract_range(range, RANGE_NUM, start, end + 1); in early_root_info_init() 189 if (!range[i].end) in early_root_info_init() 192 update_res(info, range[i].start, range[i].end - 1, in early_root_info_init() 197 memset(range, 0, sizeof(range)); in early_root_info_init() 201 add_range(range, RANGE_NUM, 0, 0, end); in early_root_info_init() 209 subtract_range(range, RANGE_NUM, 0, end); in early_root_info_init() 218 subtract_range(range, RANGE_NUM, fam10h_mmconf_start, in early_root_info_init() [all …]
|
/linux/drivers/virt/acrn/ |
H A D | ioreq.c | 118 struct acrn_ioreq_range *range; in acrn_ioreq_range_add() local 126 range = kzalloc(sizeof(*range), GFP_KERNEL); in acrn_ioreq_range_add() 127 if (!range) in acrn_ioreq_range_add() 130 range->type = type; in acrn_ioreq_range_add() 131 range->start = start; in acrn_ioreq_range_add() 132 range->end = end; in acrn_ioreq_range_add() 135 list_add(&range->list, &client->range_list); in acrn_ioreq_range_add() 151 struct acrn_ioreq_range *range; in acrn_ioreq_range_del() local 154 list_for_each_entry(range, &client->range_list, list) { in acrn_ioreq_range_del() 155 if (type == range->type && in acrn_ioreq_range_del() [all …]
|
/linux/drivers/nvdimm/ |
H A D | badrange.c | 214 struct badblocks *bb, const struct range *range) in badblocks_populate() argument 225 if (bre_end < range->start) in badblocks_populate() 227 if (bre->start > range->end) in badblocks_populate() 230 if (bre->start >= range->start) { in badblocks_populate() 234 if (bre_end <= range->end) in badblocks_populate() 237 len = range->start + range_len(range) in badblocks_populate() 239 __add_badblock_range(bb, start - range->start, len); in badblocks_populate() 246 if (bre->start < range->start) { in badblocks_populate() 249 if (bre_end < range->end) in badblocks_populate() 250 len = bre->start + bre->length - range->start; in badblocks_populate() [all …]
|
/linux/Documentation/userspace-api/media/v4l/ |
H A D | colorspaces-details.rst | 17 range. The chromaticities of the primary colors and the white reference 77 Y' is clamped to the range [0…1] and Cb and Cr are clamped to the range 82 The default quantization is limited range, but full range is possible 94 quantization is limited range. The chromaticities of the primary colors 121 Transfer function. Normally L is in the range [0…1], but for the 122 extended gamut xvYCC encoding values outside that range are allowed. 153 Y' is clamped to the range [0…1] and Cb and Cr are clamped to the range 156 The default quantization is limited range, but full range is possible 168 that are outside the range [0…1]. The resulting Y', Cb and Cr values are 169 scaled and offset according to the limited range formula: [all …]
|
/linux/lib/ |
H A D | test_hmm.c | 207 static int dmirror_do_fault(struct dmirror *dmirror, struct hmm_range *range) in dmirror_do_fault() argument 209 unsigned long *pfns = range->hmm_pfns; in dmirror_do_fault() 212 for (pfn = (range->start >> PAGE_SHIFT); in dmirror_do_fault() 213 pfn < (range->end >> PAGE_SHIFT); in dmirror_do_fault() 231 else if (WARN_ON(range->default_flags & HMM_PFN_WRITE)) in dmirror_do_fault() 258 const struct mmu_notifier_range *range, in dmirror_interval_invalidate() argument 267 if (range->event == MMU_NOTIFY_MIGRATE && in dmirror_interval_invalidate() 268 range->owner == dmirror->mdevice) in dmirror_interval_invalidate() 271 if (mmu_notifier_range_blockable(range)) in dmirror_interval_invalidate() 277 dmirror_do_update(dmirror, range in dmirror_interval_invalidate() 288 dmirror_range_fault(struct dmirror * dmirror,struct hmm_range * range) dmirror_range_fault() argument 333 struct hmm_range range = { dmirror_fault() local 1039 dmirror_mkentry(struct dmirror * dmirror,struct hmm_range * range,unsigned char * perm,unsigned long entry) dmirror_mkentry() argument 1081 dmirror_snapshot_invalidate(struct mmu_interval_notifier * mni,const struct mmu_notifier_range * range,unsigned long cur_seq) dmirror_snapshot_invalidate() argument 1108 dmirror_range_snapshot(struct dmirror * dmirror,struct hmm_range * range,unsigned char * perm) dmirror_range_snapshot() argument 1175 struct hmm_range range = { dmirror_snapshot() local [all...] |
H A D | nlattr.c | 120 struct netlink_range_validation *range) in nla_get_range_unsigned() argument 125 range->min = 0; in nla_get_range_unsigned() 129 range->max = U8_MAX; in nla_get_range_unsigned() 134 range->max = U16_MAX; in nla_get_range_unsigned() 138 range->max = U32_MAX; in nla_get_range_unsigned() 143 range->max = U64_MAX; in nla_get_range_unsigned() 153 range->min = pt->min; in nla_get_range_unsigned() 154 range->max = pt->max; in nla_get_range_unsigned() 157 *range = *pt->range; in nla_get_range_unsigned() 160 range->min = pt->min; in nla_get_range_unsigned() [all …]
|
/linux/drivers/gpu/drm/logicvc/ |
H A D | logicvc_of.c | 42 .range = { 50 .range = { 57 .range = { 8, 24 }, 79 .range = { 8, 24 }, 84 .range = { 92 .range = { 158 if (property->range[0] || property->range[1]) in logicvc_of_property_parse_u32() 159 if (value < property->range[0] || value > property->range[1]) in logicvc_of_property_parse_u32()
|
/linux/include/linux/comedi/ |
H A D | comedidev.h | 636 struct comedi_krange range[] __counted_by(length); 655 unsigned int range) in comedi_range_is_bipolar() argument 657 return s->range_table->range[range].min < 0; in comedi_range_is_bipolar() 676 unsigned int range) in comedi_range_is_unipolar() argument 678 return s->range_table->range[range].min >= 0; in comedi_range_is_unipolar() 697 unsigned int range) in comedi_range_is_external() argument 699 return !!(s->range_table->range[range].flags & RF_EXTERNAL); in comedi_range_is_external() 720 unsigned int range) in comedi_chan_range_is_bipolar() argument 722 return s->range_table_list[chan]->range[range].min < 0; in comedi_chan_range_is_bipolar() 743 unsigned int range) in comedi_chan_range_is_unipolar() argument [all …]
|
/linux/include/drm/ |
H A D | drm_gpuvm.h | 102 u64 range; member 157 u64 addr, u64 range); 159 u64 addr, u64 range); 163 static inline void drm_gpuva_init(struct drm_gpuva *va, u64 addr, u64 range, in drm_gpuva_init() argument 167 va->va.range = range; in drm_gpuva_init() 339 u64 start_offset, u64 range, 362 bool drm_gpuvm_range_valid(struct drm_gpuvm *gpuvm, u64 addr, u64 range); 363 bool drm_gpuvm_interval_empty(struct drm_gpuvm *gpuvm, u64 addr, u64 range); 562 u64 addr, u64 range, 572 u64 addr, u64 range); [all …]
|
/linux/drivers/comedi/drivers/ |
H A D | jr3_pci.c | 94 char _reserved[offsetof(struct comedi_lrange, range[1])]; 112 union jr3_pci_single_range range[9]; member 508 union jr3_pci_single_range *r = spriv->range; in jr3_pci_poll_subdevice() 511 r[0].l.range[0].min = -get_s16(&fs->fx) * 1000; in jr3_pci_poll_subdevice() 512 r[0].l.range[0].max = get_s16(&fs->fx) * 1000; in jr3_pci_poll_subdevice() 513 r[1].l.range[0].min = -get_s16(&fs->fy) * 1000; in jr3_pci_poll_subdevice() 514 r[1].l.range[0].max = get_s16(&fs->fy) * 1000; in jr3_pci_poll_subdevice() 515 r[2].l.range[0].min = -get_s16(&fs->fz) * 1000; in jr3_pci_poll_subdevice() 516 r[2].l.range[0].max = get_s16(&fs->fz) * 1000; in jr3_pci_poll_subdevice() 517 r[3].l.range[0].min = -get_s16(&fs->mx) * 100; in jr3_pci_poll_subdevice() [all …]
|
/linux/drivers/infiniband/core/ |
H A D | umem_odp.c | 355 struct hmm_range range = {}; in ib_umem_odp_map_dma_and_lock() local 378 range.notifier = &umem_odp->notifier; in ib_umem_odp_map_dma_and_lock() 379 range.start = ALIGN_DOWN(user_virt, 1UL << page_shift); in ib_umem_odp_map_dma_and_lock() 380 range.end = ALIGN(user_virt + bcnt, 1UL << page_shift); in ib_umem_odp_map_dma_and_lock() 381 pfn_start_idx = (range.start - ib_umem_start(umem_odp)) >> PAGE_SHIFT; in ib_umem_odp_map_dma_and_lock() 382 num_pfns = (range.end - range.start) >> PAGE_SHIFT; in ib_umem_odp_map_dma_and_lock() 384 range.default_flags = HMM_PFN_REQ_FAULT; in ib_umem_odp_map_dma_and_lock() 387 range.default_flags |= HMM_PFN_REQ_WRITE; in ib_umem_odp_map_dma_and_lock() 390 range.hmm_pfns = &(umem_odp->pfn_list[pfn_start_idx]); in ib_umem_odp_map_dma_and_lock() 394 current_seq = range.notifier_seq = in ib_umem_odp_map_dma_and_lock() [all …]
|
/linux/drivers/gpu/drm/ |
H A D | drm_gpuvm.c | 871 #define GPUVA_LAST(node) ((node)->va.addr + (node)->va.range - 1) 885 drm_gpuvm_check_overflow(u64 addr, u64 range) in drm_gpuvm_check_overflow() argument 889 return check_add_overflow(addr, range, &end); in drm_gpuvm_check_overflow() 893 drm_gpuvm_warn_check_overflow(struct drm_gpuvm *gpuvm, u64 addr, u64 range) in drm_gpuvm_warn_check_overflow() argument 895 return drm_WARN(gpuvm->drm, drm_gpuvm_check_overflow(addr, range), in drm_gpuvm_warn_check_overflow() 900 drm_gpuvm_in_mm_range(struct drm_gpuvm *gpuvm, u64 addr, u64 range) in drm_gpuvm_in_mm_range() argument 902 u64 end = addr + range; in drm_gpuvm_in_mm_range() 910 drm_gpuvm_in_kernel_node(struct drm_gpuvm *gpuvm, u64 addr, u64 range) in drm_gpuvm_in_kernel_node() argument 912 u64 end = addr + range; in drm_gpuvm_in_kernel_node() 914 u64 krange = gpuvm->kernel_alloc_node.va.range; in drm_gpuvm_in_kernel_node() [all …]
|