Home
last modified time | relevance | path

Searched full:range (Results 1 – 25 of 4552) sorted by relevance

12345678910>>...183

/linux/kernel/
H A Drange.c3 * Range add and subtract
10 #include <linux/range.h>
12 int add_range(struct range *range, int az, int nr_range, u64 start, u64 end) in add_range() argument
21 range[nr_range].start = start; in add_range()
22 range[nr_range].end = end; in add_range()
29 int add_range_with_merge(struct range *range, int az, int nr_range, in add_range_with_merge() argument
41 if (!range[i].end) in add_range_with_merge()
44 common_start = max(range[i].start, start); in add_range_with_merge()
45 common_end = min(range[i].end, end); in add_range_with_merge()
50 start = min(range[i].start, start); in add_range_with_merge()
[all …]
/linux/Documentation/gpu/
H A Dkms-properties.csv7 ,,“left margin”,RANGE,"Min=0, Max=100",Connector,TBD
8 ,,“right margin”,RANGE,"Min=0, Max=100",Connector,TBD
9 ,,“top margin”,RANGE,"Min=0, Max=100",Connector,TBD
10 ,,“bottom margin”,RANGE,"Min=0, Max=100",Connector,TBD
11 ,,“brightness”,RANGE,"Min=0, Max=100",Connector,TBD
12 ,,“contrast”,RANGE,"Min=0, Max=100",Connector,TBD
13 ,,“flicker reduction”,RANGE,"Min=0, Max=100",Connector,TBD
14 ,,“overscan”,RANGE,"Min=0, Max=100",Connector,TBD
15 ,,“saturation”,RANGE,"Min=0, Max=100",Connector,TBD
16 ,,“hue”,RANGE,"Min=0, Max=100",Connector,TBD
[all …]
/linux/drivers/gpu/drm/
H A Ddrm_gpusvm.c46 * space is chosen for the range size. Ranges are expected to be
53 * range allocation, notifier allocation, and invalidations.
74 * optionally migrate the range to device memory, and create GPU bindings.
92 * range RB tree and list, as well as the range's DMA mappings and sequence
94 * except for the recheck range's pages being valid
115 * being that a subset of the range still has CPU and GPU mappings. If the
116 * backing store for the range is in device memory, a subset of the backing
117 * store has references. One option would be to split the range and device
119 * complicated. Given that partial unmappings are rare and driver-defined range
122 * With no support for range splitting, upon partial unmapping of a range, the
[all …]
/linux/drivers/soc/ti/
H A Dknav_qmss_acc.c20 #define knav_range_offset_to_inst(kdev, range, q) \ argument
21 (range->queue_base_inst + (q << kdev->inst_shift))
23 static void __knav_acc_notify(struct knav_range_info *range, in __knav_acc_notify() argument
26 struct knav_device *kdev = range->kdev; in __knav_acc_notify()
30 range_base = kdev->base_id + range->queue_base; in __knav_acc_notify()
32 if (range->flags & RANGE_MULTI_QUEUE) { in __knav_acc_notify()
33 for (queue = 0; queue < range->num_queues; queue++) { in __knav_acc_notify()
34 inst = knav_range_offset_to_inst(kdev, range, in __knav_acc_notify()
44 queue = acc->channel - range->acc_info.start_channel; in __knav_acc_notify()
45 inst = knav_range_offset_to_inst(kdev, range, queue); in __knav_acc_notify()
[all …]
/linux/drivers/gpu/drm/xe/
H A Dxe_svm.h23 /** struct xe_svm_range - SVM range */
28 * @garbage_collector_link: Link into VM's garbage collect SVM range
33 * @tile_present: Tile mask of binding is present for this range.
39 * range. Protected by GPU SVM notifier lock.
45 * xe_svm_range_pages_valid() - SVM range pages valid
46 * @range: SVM range
48 * Return: True if SVM range pages are valid, False otherwise
50 static inline bool xe_svm_range_pages_valid(struct xe_svm_range *range) in xe_svm_range_pages_valid() argument
52 return drm_gpusvm_range_pages_valid(range->base.gpusvm, &range->base); in xe_svm_range_pages_valid()
71 void xe_svm_range_debug(struct xe_svm_range *range, const char *operation);
[all …]
H A Dxe_svm.c21 static bool xe_svm_range_in_vram(struct xe_svm_range *range) in xe_svm_range_in_vram() argument
24 * Advisory only check whether the range is currently backed by VRAM in xe_svm_range_in_vram()
30 .__flags = READ_ONCE(range->base.flags.__flags), in xe_svm_range_in_vram()
36 static bool xe_svm_range_has_vram_binding(struct xe_svm_range *range) in xe_svm_range_has_vram_binding() argument
39 return xe_svm_range_in_vram(range) && range->tile_present; in xe_svm_range_has_vram_binding()
64 void xe_svm_range_debug(struct xe_svm_range *range, const char *operation) in xe_svm_range_debug() argument
66 range_debug(range, operation); in xe_svm_range_debug()
77 struct xe_svm_range *range; in xe_svm_range_alloc() local
79 range = kzalloc(sizeof(*range), GFP_KERNEL); in xe_svm_range_alloc()
80 if (!range) in xe_svm_range_alloc()
[all …]
H A Dxe_range_fence.c44 * xe_range_fence_insert() - range fence insert
45 * @tree: range fence tree to insert intoi
46 * @rfence: range fence
47 * @ops: range fence ops
48 * @start: start address of range fence
49 * @last: last address of range fence
50 * @fence: dma fence which signals range fence can be removed + freed
107 * xe_range_fence_tree_init() - Init range fence tree
108 * @tree: range fence tree
116 * xe_range_fence_tree_fini() - Fini range fence tree
[all …]
/linux/arch/s390/include/asm/
H A Dphysmem_info.h90 * for_each_physmem_usable_range - early online memory range iterator
92 * @p_start: ptr to unsigned long for start address of the range
93 * @p_end: ptr to unsigned long for end address of the range
138 #define for_each_physmem_reserved_type_range(t, range, p_start, p_end) \ argument
139 for (range = &physmem_info.reserved[t], *p_start = range->start, *p_end = range->end; \
140 range && range->end; range = range->chain ? __va(range->chain) : NULL, \
141 *p_start = range ? range->start : 0, *p_end = range ? range->end : 0)
144 struct reserved_range *range) in __physmem_reserved_next() argument
146 if (!range) { in __physmem_reserved_next()
147 range = &physmem_info.reserved[*t]; in __physmem_reserved_next()
[all …]
/linux/include/drm/
H A Ddrm_gpusvm.h48 * @range_alloc: Allocate a GPU SVM range (optional)
51 * Allocate a GPU SVM range.
53 * Return: Pointer to the allocated GPU SVM range on success, NULL on failure.
58 * @range_free: Free a GPU SVM range (optional)
59 * @range: Pointer to the GPU SVM range to be freed
61 * Free a GPU SVM range.
63 void (*range_free)(struct drm_gpusvm_range *range);
71 * Invalidate the GPU page tables. It can safely walk the notifier range
109 * struct drm_gpusvm_range_flags - Structure representing a GPU SVM range flags
111 * @migrate_devmem: Flag indicating whether the range can be migrated to device memory
[all …]
/linux/tools/testing/selftests/net/
H A Dip_local_port_range.c28 static void unpack_port_range(__u32 range, __u16 *lo, __u16 *hi) in unpack_port_range() argument
30 *lo = range & 0xffff; in unpack_port_range()
31 *hi = range >> 16; in unpack_port_range()
104 static int get_ip_local_port_range(int fd, __u32 *range) in get_ip_local_port_range() argument
115 *range = val; in get_ip_local_port_range()
199 /* Empty range: low port > high port */ in TEST_F()
221 { 30000, 39999 }, /* socket range below netns range */ in TEST_F()
222 { 50000, 59999 }, /* socket range above netns range */ in TEST_F()
228 * that the range wasn't clamped to a single port from in TEST_F()
229 * the netns range. That is [40000, 40000] or [49999, in TEST_F()
[all …]
/linux/include/linux/
H A Drange.h6 struct range { struct
11 static inline u64 range_len(const struct range *range) in range_len() argument
13 return range->end - range->start + 1; in range_len()
17 static inline bool range_contains(const struct range *r1, in range_contains()
18 const struct range *r2) in range_contains()
24 static inline bool range_overlaps(const struct range *r1, in range_overlaps()
25 const struct range *r2) in range_overlaps()
30 int add_range(struct range *range, int az, int nr_range,
34 int add_range_with_merge(struct range *range, int az, int nr_range,
37 void subtract_range(struct range *range, int az, u64 start, u64 end);
[all …]
/linux/block/
H A Dbadblocks.c23 * When the caller of badblocks_set() wants to set a range of bad blocks, the
24 * setting range can be acked or unacked. And the setting range may merge,
25 * overwrite, skip the overlapped already set range, depends on who they are
27 * more complicated when the setting range covers multiple already set bad block
28 * ranges, with restrictions of maximum length of each bad range and the bad
32 * for setting a large range of bad blocks, we can handle it by dividing the
33 * large range into smaller ones when encounter overlap, max range length or
34 * bad table full conditions. Every time only a smaller piece of the bad range
39 * When setting a range of bad blocks to the bad table, the simplified situations
41 * prefix E, and the setting bad blocks range is naming with prefix S)
[all …]
/linux/mm/
H A Dmemremap.c41 static void pgmap_array_delete(struct range *range)
43 xa_store_range(&pgmap_array, PHYS_PFN(range->start), PHYS_PFN(range->end),
50 struct range *range = &pgmap->ranges[range_id]; in devmap_managed_enable_put()
51 unsigned long pfn = PHYS_PFN(range->start);
63 struct range *range = &pgmap->ranges[i]; in devmap_managed_enable_put()
65 if (pfn >= PHYS_PFN(range
66 pgmap_array_delete(struct range * range) pgmap_array_delete() argument
75 struct range *range = &pgmap->ranges[range_id]; pfn_first() local
88 struct range *range = &pgmap->ranges[i]; pgmap_pfn_valid() local
100 const struct range *range = &pgmap->ranges[range_id]; pfn_end() local
113 struct range *range = &pgmap->ranges[range_id]; pageunmap_range() local
174 struct range *range = &pgmap->ranges[range_id]; pagemap_range() local
[all...]
H A Dhmm.c34 struct hmm_range *range; member
51 struct hmm_range *range, unsigned long cpu_flags) in hmm_pfns_fill() argument
53 unsigned long i = (addr - range->start) >> PAGE_SHIFT; in hmm_pfns_fill()
56 range->hmm_pfns[i] &= HMM_PFN_INOUT_FLAGS; in hmm_pfns_fill()
57 range->hmm_pfns[i] |= cpu_flags; in hmm_pfns_fill()
63 * hmm_vma_fault() - fault in a range lacking valid pmd or pte(s)
64 * @addr: range virtual start address (inclusive)
65 * @end: range virtual end address (exclusive)
71 * or whenever there is no page directory covering the virtual address range.
100 struct hmm_range *range in hmm_pte_need_fault() local
135 struct hmm_range *range = hmm_vma_walk->range; hmm_range_need_fault() local
161 struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk_hole() local
186 pmd_to_hmm_pfn_flags(struct hmm_range * range,pmd_t pmd) pmd_to_hmm_pfn_flags() argument
202 struct hmm_range *range = hmm_vma_walk->range; hmm_vma_handle_pmd() local
227 pte_to_hmm_pfn_flags(struct hmm_range * range,pte_t pte) pte_to_hmm_pfn_flags() argument
240 struct hmm_range *range = hmm_vma_walk->range; hmm_vma_handle_pte() local
338 struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk_pmd() local
413 pud_to_hmm_pfn_flags(struct hmm_range * range,pud_t pud) pud_to_hmm_pfn_flags() argument
427 struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk_pud() local
488 struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk_hugetlb_entry() local
540 struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk_test() local
598 hmm_range_fault(struct hmm_range * range) hmm_range_fault() argument
[all...]
H A Dmadvise.c74 * The range over which the behaviour is currently being applied. If
77 struct madvise_behavior_range range; member
156 struct madvise_behavior_range *range = &madv_behavior->range; in madvise_update_vma() local
159 VMA_ITERATOR(vmi, madv_behavior->mm, range->start); in madvise_update_vma()
167 range->start, range->end, anon_name); in madvise_update_vma()
170 range->start, range->end, new_flags); in madvise_update_vma()
288 unsigned long start = madv_behavior->range.start; in madvise_willneed()
289 unsigned long end = madv_behavior->range.end; in madvise_willneed()
485 * fully mapped within the range we are operating on. Otherwise in madvise_cold_or_pageout_pte_range()
488 * next pte in the range. in madvise_cold_or_pageout_pte_range()
[all …]
/linux/lib/
H A Dlinear_ranges.c3 * helpers to map values in a linear range to range index
18 * linear_range_values_in_range - return the amount of values in a range
19 * @r: pointer to linear range where values are counted
21 * Compute the amount of values in range pointed by @r. Note, values can
22 * be all equal - range with selectors 0,...,2 with step 0 still contains
25 * Return: the amount of values in range pointed by @r
41 * be all equal - range with selectors 0,...,2 with step 0 still contains
65 * linear_range_get_max_value - return the largest value in a range
66 * @r: pointer to linear range where value is looked from
68 * Return: the largest value in the given range
[all …]
/linux/drivers/of/
H A Daddress.c33 u64 (*map)(__be32 *addr, const __be32 *range,
53 static u64 of_bus_default_map(__be32 *addr, const __be32 *range, in of_bus_default_map() argument
58 cp = of_read_number(range + fna, na - fna); in of_bus_default_map()
59 s = of_read_number(range + na + pna, ns); in of_bus_default_map()
91 static u64 of_bus_default_flags_map(__be32 *addr, const __be32 *range, int na, in of_bus_default_flags_map() argument
95 if (*addr != *range) in of_bus_default_flags_map()
98 return of_bus_default_map(addr, range, na, ns, pna, fna); in of_bus_default_flags_map()
171 static u64 of_bus_pci_map(__be32 *addr, const __be32 *range, int na, int ns, in of_bus_pci_map() argument
177 rf = of_bus_pci_get_flags(range); in of_bus_pci_map()
183 return of_bus_default_map(addr, range, na, ns, pna, fna); in of_bus_pci_map()
[all …]
/linux/arch/mips/loongson64/
H A Dinit.c156 struct logic_pio_hwaddr *range; in add_legacy_isa_io() local
159 range = kzalloc(sizeof(*range), GFP_ATOMIC); in add_legacy_isa_io()
160 if (!range) in add_legacy_isa_io()
163 range->fwnode = fwnode; in add_legacy_isa_io()
164 range->size = size = round_up(size, PAGE_SIZE); in add_legacy_isa_io()
165 range->hw_start = hw_start; in add_legacy_isa_io()
166 range->flags = LOGIC_PIO_CPU_MMIO; in add_legacy_isa_io()
168 ret = logic_pio_register_range(range); in add_legacy_isa_io()
170 kfree(range); in add_legacy_isa_io()
175 if (range->io_start != 0) { in add_legacy_isa_io()
[all …]
/linux/arch/powerpc/kexec/
H A Dranges.c40 sizeof(struct range)); in get_max_nr_ranges()
58 (mem_rngs->max_nr_ranges * sizeof(struct range))); in get_mem_rngs_size()
68 * __add_mem_range - add a memory range to memory ranges list.
69 * @mem_ranges: Range list to add the memory range to.
70 * @base: Base address of the range to add.
71 * @size: Size of the memory range to add.
89 pr_debug("Added memory range [%#016llx - %#016llx] at index %d\n", in __add_mem_range()
97 * @mem_rngs: Range list to merge.
99 * Assumes a sorted range list.
105 struct range *ranges; in __merge_memory_ranges()
[all …]
/linux/tools/testing/selftests/bpf/prog_tests/
H A Dreg_bounds.c215 * GENERIC RANGE STRUCT AND OPERATIONS
218 struct range { struct
222 static void snprintf_range(enum num_t t, struct strbuf *sb, struct range x) in snprintf_range() argument
234 static void print_range(enum num_t t, struct range x, const char *sfx) in print_range()
242 static const struct range unkn[] = {
249 static struct range unkn_subreg(enum num_t t) in unkn_subreg()
260 static struct range range(enum num_t t, u64 a, u64 b) in range() function
263 case U64: return (struct range){ (u64)a, (u64)b }; in range()
264 case U32: return (struct range){ (u32)a, (u32)b }; in range()
265 case S64: return (struct range){ (s64)a, (s64)b }; in range()
[all …]
/linux/drivers/net/wireless/ti/wl12xx/
H A Dconf.h19 * Range: s8
26 * Range: s8
41 * Range: 0 - 255 (ms)
50 * Range: 0 - 255 (ms)
59 * Range: 0 - 255 (ms)
68 * Range: 0 - 255 (ms)
77 * Range: 0 - 255 (ms)
86 * Range: 0 - 255 (ms)
112 * Range: 0 - 255 (%)
120 * Range: 0 - 255 (%)
[all …]
/linux/drivers/net/wireless/ti/wlcore/
H A Dconf.h117 * Range: 0 - 0xFFFFFFFF
132 * Range: 0 - 200000
139 * Range: 0 - 200000
147 * Range: 0 - 4096
155 * Range: ENABLE_ENERGY_D == 0x140A
164 * Range: u16
172 * Range: u16
179 * Range: 1 - 100
186 * Range: RX_QUEUE_TYPE_RX_LOW_PRIORITY, RX_QUEUE_TYPE_RX_HIGH_PRIORITY,
252 * Range: CONF_HW_BIT_RATE_* bit mask
[all …]
/linux/drivers/gpu/ipu-v3/
H A Dipu-ic-csc.c29 * RGB full-range to RGB limited-range
46 * RGB limited-range to RGB full-range
63 * YUV full-range to YUV limited-range
81 * YUV limited-range to YUV full-range
112 * BT.601 RGB full-range to YUV full-range
128 /* BT.601 RGB full-range to YUV limited-range */
140 /* BT.601 RGB limited-range to YUV full-range */
151 /* BT.601 RGB limited-range to YUV limited-range */
164 * BT.601 YUV full-range to RGB full-range
186 /* BT.601 YUV full-range to RGB limited-range */
[all …]
/linux/drivers/pci/hotplug/
H A Dibmphp_res.c116 /* need to insert our range */ in alloc_bus_range()
164 * 2. If cannot allocate out of PFMem range, allocate from Mem ranges. PFmemFromMem
294 * assign a -1 and then update once the range in ibmphp_rsrc_init()
336 * range actually appears... in ibmphp_rsrc_init()
361 * This function adds a range into a sorted list of ranges per bus for a particular
362 * range type, it then calls another routine to update the range numbers on the
365 * Input: type of the resource, range to add, current bus
366 * Output: 0 or -1, bus and range ptrs
368 static int add_bus_range(int type, struct range_node *range, struct bus_node *bus_cur) in add_bus_range() argument
392 if (range->start < range_cur->start) in add_bus_range()
[all …]
/linux/net/netfilter/
H A Dnf_nat_core.c404 const struct nf_nat_range2 *range) in nf_nat_inet_in_range() argument
407 return ntohl(t->src.u3.ip) >= ntohl(range->min_addr.ip) && in nf_nat_inet_in_range()
408 ntohl(t->src.u3.ip) <= ntohl(range->max_addr.ip); in nf_nat_inet_in_range()
410 return ipv6_addr_cmp(&t->src.u3.in6, &range->min_addr.in6) >= 0 && in nf_nat_inet_in_range()
411 ipv6_addr_cmp(&t->src.u3.in6, &range->max_addr.in6) <= 0; in nf_nat_inet_in_range()
445 * that meet the constraints of range.
448 const struct nf_nat_range2 *range) in nf_in_range() argument
451 * range specified, otherwise let this drag us onto a new src IP. in nf_in_range()
453 if (range->flags & NF_NAT_RANGE_MAP_IPS && in nf_in_range()
454 !nf_nat_inet_in_range(tuple, range)) in nf_in_range()
[all …]

12345678910>>...183