| /linux/scripts/ |
| H A D | decode_stacktrace.sh | 287 # We need to know the index of the last element before we 289 local last=$(( ${#words[@]} - 1 )) 304 if [[ ${words[$last]} =~ \([A-Z]*\) ]]; then 305 info_str=${words[$last]} 306 unset words[$last] spaces[$last] 307 last=$(( $last - 1 )) 312 if [[ ${words[$last]} =~ ^[0-9a-f]+\] ]]; then 313 words[$last [all...] |
| /linux/drivers/infiniband/hw/usnic/ |
| H A D | usnic_uiom_interval_tree.c | 43 #define LAST(node) ((node)->last) 70 usnic_uiom_interval_node_alloc(long int start, long int last, int ref_cnt, in usnic_uiom_interval_node_alloc() argument 79 interval->last = last; in usnic_uiom_interval_node_alloc() 105 unsigned long start, unsigned long last, in find_intervals_intersection_sorted() argument 112 for (node = usnic_uiom_interval_tree_iter_first(root, start, last); in find_intervals_intersection_sorted() 114 node = usnic_uiom_interval_tree_iter_next(node, start, last)) in find_intervals_intersection_sorted() 120 int usnic_uiom_get_intervals_diff(unsigned long start, unsigned long last, in usnic_uiom_get_intervals_diff() argument 132 find_intervals_intersection_sorted(root, start, last, in usnic_uiom_get_intervals_diff() 148 if (pivot > interval->last) { in usnic_uiom_get_intervals_diff() 150 } else if (pivot <= interval->last && in usnic_uiom_get_intervals_diff() [all …]
|
| /linux/fs/xfs/scrub/ |
| H A D | bitmap.c | 51 uint64_t last); 55 uint64_t last); 78 uint64_t last = start + len - 1; local 80 while ((bn = xbitmap64_tree_iter_first(&bitmap->xb_root, start, last))) { 81 if (bn->bn_start < start && bn->bn_last > last) { 94 new_bn->bn_start = last + 1; 102 } else if (bn->bn_last > last) { 105 bn->bn_start = last + 1; 127 uint64_t last = start + len - 1; in xbitmap64_set() local 131 left = xbitmap64_tree_iter_first(&bitmap->xb_root, start, last); in xbitmap64_set() [all …]
|
| /linux/rust/syn/ |
| H A D | punctuated.rs | 51 last: Option<Box<T>>, field 59 last: None, in new() 66 self.inner.len() == 0 && self.last.is_none() in is_empty() 74 self.inner.len() + if self.last.is_some() { 1 } else { 0 } in len() 88 pub fn last(&self) -> Option<&T> { in last() method 102 self.last.as_deref() in get() 114 self.last.as_deref_mut() in get_mut() 125 last: self.last.as_ref().map(Box::as_ref).into_iter(), in iter() 136 last: self.last.as_mut().map(Box::as_mut).into_iter(), in iter_mut() 146 last: self.last.as_ref().map(Box::as_ref).into_iter(), in pairs() [all …]
|
| /linux/arch/mips/sibyte/common/ |
| H A D | cfe_console.c | 16 int i, last, written; in cfe_console_write() local 18 for (i=0, last=0; i<count; i++) { in cfe_console_write() 24 written = cfe_write(cfe_cons_handle, &str[last], i-last); in cfe_console_write() 27 last += written; in cfe_console_write() 28 } while (last < i); in cfe_console_write() 33 if (last != count) { in cfe_console_write() 35 written = cfe_write(cfe_cons_handle, &str[last], count-last); in cfe_console_write() 38 last += written; in cfe_console_write() 39 } while (last < count); in cfe_console_write()
|
| /linux/drivers/vhost/ |
| H A D | iotlb.c | 17 #define LAST(map) ((map)->last) 51 u64 start, u64 last, in vhost_iotlb_add_range_ctx() argument 57 if (last < start) in vhost_iotlb_add_range_ctx() 63 if (start == 0 && last == ULONG_MAX) { in vhost_iotlb_add_range_ctx() 64 u64 mid = last / 2; in vhost_iotlb_add_range_ctx() 87 map->size = last - start + 1; in vhost_iotlb_add_range_ctx() 88 map->last = last; in vhost_iotlb_add_range_ctx() 104 u64 start, u64 last, in vhost_iotlb_add_range() argument 107 return vhost_iotlb_add_range_ctx(iotlb, start, last, in vhost_iotlb_add_range() 118 void vhost_iotlb_del_range(struct vhost_iotlb *iotlb, u64 start, u64 last) in vhost_iotlb_del_range() argument [all …]
|
| /linux/tools/testing/selftests/vfio/lib/ |
| H A D | iova_allocator.c | 58 iova_t iova, last; in iova_allocator_alloc() local 67 if (check_add_overflow(iova, size - 1, &last) || in iova_allocator_alloc() 68 last > range->last) in iova_allocator_alloc() 72 iova = last & ~(size - 1); in iova_allocator_alloc() 75 if (check_add_overflow(iova, size - 1, &last) || in iova_allocator_alloc() 76 last > range->last) in iova_allocator_alloc() 79 if (last == range->last) { in iova_allocator_alloc() 83 allocator->range_offset = last - range->start + 1; in iova_allocator_alloc()
|
| /linux/lib/ |
| H A D | interval_tree_test.c | 33 search(struct rb_root_cached *root, unsigned long start, unsigned long last) in search() argument 38 for (node = interval_tree_iter_first(root, start, last); node; in search() 39 node = interval_tree_iter_next(node, start, last)) in search() 53 nodes[i].last = b; in init() 111 unsigned long last = search_all ? max_endpoint : queries[j]; in search_check() local 113 results += search(&root, start, last); in search_check() 133 unsigned long start, last; in intersection_range_check() local 164 last = ULONG_MAX; in intersection_range_check() 166 last = (prandom_u32_state(&rnd) >> 4) % max_endpoint; in intersection_range_check() 167 start = (prandom_u32_state(&rnd) >> 4) % last; in intersection_range_check() [all …]
|
| H A D | test_maple_tree.c | 393 unsigned long last = 0, index = 0; in check_find() local 573 last = mas.last; in check_find() 578 mas.last = ULONG_MAX; in check_find() 588 MT_BUG_ON(mt, last != mas.last); in check_find() 593 mas.last = ULONG_MAX; in check_find() 785 pr_debug("Found %lu %lu\n", mas.index, mas.last); in check_alloc_rev_range() 789 MT_BUG_ON(mt, mas.last + 1 != (holes[i+1] >> 12)); in check_alloc_rev_range() 1523 MT_BUG_ON(mt, mas.last != ULONG_MAX); in check_root_expand() 1555 MT_BUG_ON(mt, mas.last != ULONG_MAX); in check_root_expand() 1560 MT_BUG_ON(mt, mas.last != 0); in check_root_expand() [all …]
|
| H A D | klist.c | 336 struct klist_node *last = i->i_cur; in klist_prev() local 342 if (last) { in klist_prev() 343 prev = to_klist_node(last->n_node.prev); in klist_prev() 344 if (!klist_dec_and_del(last)) in klist_prev() 361 if (put && last) in klist_prev() 362 put(last); in klist_prev() 378 struct klist_node *last = i->i_cur; in klist_next() local 384 if (last) { in klist_next() 385 next = to_klist_node(last->n_node.next); in klist_next() 386 if (!klist_dec_and_del(last)) in klist_next() [all …]
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_svm.c | 78 svm_range_check_vm(struct kfd_process *p, uint64_t start, uint64_t last, 96 prange, prange->start, prange->last); in svm_range_unlink() 105 if (prange->it_node.start != 0 && prange->it_node.last != 0) in svm_range_unlink() 113 prange, prange->start, prange->last); in svm_range_add_notifier_locked() 132 prange, prange->start, prange->last); in svm_range_add_to_svms() 136 prange->it_node.last = prange->last; in svm_range_add_to_svms() 145 prange->notifier.interval_tree.last >> PAGE_SHIFT); in svm_range_remove_notifier() 148 prange->notifier.interval_tree.last != 0) in svm_range_remove_notifier() 282 uint64_t size = (prange->last - prange->start + 1) << PAGE_SHIFT; in svm_range_free() 287 prange->start, prange->last); in svm_range_free() [all …]
|
| /linux/include/trace/events/ |
| H A D | maple_tree.h | 24 __field(unsigned long, last) 33 __entry->last = mas->last; 43 (unsigned long) __entry->last 57 __field(unsigned long, last) 66 __entry->last = mas->last; 76 (unsigned long) __entry->last 92 __field(unsigned long, last) 103 __entry->last = mas->last; 115 (unsigned long) __entry->last,
|
| /linux/tools/perf/util/ |
| H A D | ordered-events.c | 20 struct ordered_event *last = oe->last; in queue_event() local 25 oe->last = new; in queue_event() 29 if (!last) { in queue_event() 36 * last event might point to some random place in the list as it's in queue_event() 37 * the last queued event. We expect that the new event is close to in queue_event() 40 if (last->timestamp <= timestamp) { in queue_event() 41 while (last->timestamp <= timestamp) { in queue_event() 42 p = last->list.next; in queue_event() 48 last in queue_event() 294 struct ordered_event *first, *last; __ordered_events__flush() local [all...] |
| /linux/drivers/video/fbdev/core/ |
| H A D | fb_fillrect.h | 111 unsigned long first, last; in bitfill() local 115 last = ~fb_pixel_mask(end & (BITS_PER_LONG-1), pattern->reverse); in bitfill() 118 last = last ? (last & first) : first; in bitfill() 120 if (last == ~0UL) in bitfill() 122 else if (last) in bitfill() 123 fb_modify_offset(first, last, 0, dst); in bitfill() 147 if (last) in bitfill() 148 fb_modify_offset(get(pattern), last, offset, dst); in bitfill() 159 unsigned long first, last; in bitinvert() local 164 last = ~fb_pixel_mask(end & (BITS_PER_LONG-1), pattern->reverse); in bitinvert() [all …]
|
| /linux/drivers/base/regmap/ |
| H A D | regcache-maple.c | 44 unsigned long index, last; in regcache_maple_write() local 60 last = reg; in regcache_maple_write() 65 lower_sz = (mas.last - mas.index + 1) * sizeof(unsigned long); in regcache_maple_write() 70 last = mas.last; in regcache_maple_write() 71 upper_sz = (mas.last - mas.index + 1) * sizeof(unsigned long); in regcache_maple_write() 76 entry = kmalloc_array(last - index + 1, sizeof(*entry), map->alloc_flags); in regcache_maple_write() 93 mas_set_range(&mas, index, last); in regcache_maple_write() 145 if (mas.last > max) { in regcache_maple_drop() 147 upper_last = mas.last; in regcache_maple_drop() 150 mas.last - max, sizeof(*upper), in regcache_maple_drop() [all …]
|
| /linux/crypto/asymmetric_keys/ |
| H A D | pkcs7_trust.c | 27 struct x509_certificate *x509, *last = NULL, *p; in pkcs7_validate_trust_one() local 74 last = x509; in pkcs7_validate_trust_one() 75 sig = last->sig; in pkcs7_validate_trust_one() 81 if (last && (last->sig->auth_ids[0] || last->sig->auth_ids[1])) { in pkcs7_validate_trust_one() 83 last->sig->auth_ids[0], in pkcs7_validate_trust_one() 84 last->sig->auth_ids[1], in pkcs7_validate_trust_one() 87 x509 = last; in pkcs7_validate_trust_one()
|
| /linux/net/netfilter/ |
| H A D | nft_set_pipapo_avx2.c | 154 union nft_pipapo_map_bucket *mt, bool last) in nft_pipapo_avx2_refill() argument 164 if (last) \ in nft_pipapo_avx2_refill() 217 bool first, bool last) in nft_pipapo_avx2_lookup_4b_2() argument 243 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_2() 244 if (last) in nft_pipapo_avx2_lookup_4b_2() 280 bool first, bool last) in nft_pipapo_avx2_lookup_4b_4() argument 320 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_4() 321 if (last) in nft_pipapo_avx2_lookup_4b_4() 357 bool first, bool last) in nft_pipapo_avx2_lookup_4b_8() argument 415 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_8() [all …]
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | frwr_ops.c | 569 struct ib_send_wr *first, **prev, *last; in frwr_unmap_sync() local 586 last = &mr->mr_invwr; in frwr_unmap_sync() 587 last->next = NULL; in frwr_unmap_sync() 588 last->wr_cqe = &mr->mr_cqe; in frwr_unmap_sync() 589 last->sg_list = NULL; in frwr_unmap_sync() 590 last->num_sge = 0; in frwr_unmap_sync() 591 last->opcode = IB_WR_LOCAL_INV; in frwr_unmap_sync() 592 last->send_flags = IB_SEND_SIGNALED; in frwr_unmap_sync() 593 last->ex.invalidate_rkey = mr->mr_handle; in frwr_unmap_sync() 595 last->wr_cqe->done = frwr_wc_localinv; in frwr_unmap_sync() [all …]
|
| /linux/drivers/memory/tegra/ |
| H A D | tegra210-emc-cc-r21021.c | 136 struct tegra210_emc_timing *curr = emc->last; in tegra210_emc_get_clktree_delay() 183 struct tegra210_emc_timing *last, in periodic_compensation_handler() argument 198 if (last->periodic_training && in periodic_compensation_handler() 207 __COPY_EMA(next, last, idx); in periodic_compensation_handler() 258 struct tegra210_emc_timing *last = emc->last; in tegra210_emc_r21021_periodic_compensation() local 261 if (last->periodic_training) { in tegra210_emc_r21021_periodic_compensation() 304 last, last)) { in tegra210_emc_r21021_periodic_compensation() 310 value = tegra210_emc_compensate(last, list[i]); in tegra210_emc_r21021_periodic_compensation() 359 struct tegra210_emc_timing *fake, *last = emc->last, *next = emc->next; in tegra210_emc_r21021_set_clock() local 379 fake = tegra210_emc_find_timing(emc, last->rate * 1000UL); in tegra210_emc_r21021_set_clock() [all …]
|
| /linux/drivers/infiniband/hw/cxgb4/ |
| H A D | id_table.c | 51 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in c4iw_id_alloc() 57 alloc->last += get_random_u32_below(RANDOM_SKIP); in c4iw_id_alloc() 59 alloc->last = obj + 1; in c4iw_id_alloc() 60 if (alloc->last >= alloc->max) in c4iw_id_alloc() 61 alloc->last = 0; in c4iw_id_alloc() 88 alloc->last = get_random_u32_below(RANDOM_SKIP); in c4iw_id_table_alloc() 90 alloc->last = 0; in c4iw_id_table_alloc()
|
| /linux/drivers/staging/media/ipu3/ |
| H A D | ipu3-css-pool.c | 49 pool->last = IPU3_CSS_POOL_SIZE; in imgu_css_pool_init() 64 u32 n = (pool->last + 1) % IPU3_CSS_POOL_SIZE; in imgu_css_pool_get() 67 pool->last = n; in imgu_css_pool_get() 75 pool->entry[pool->last].valid = false; in imgu_css_pool_put() 76 pool->last = (pool->last + IPU3_CSS_POOL_SIZE - 1) % IPU3_CSS_POOL_SIZE; in imgu_css_pool_put() 92 int i = (pool->last + IPU3_CSS_POOL_SIZE - n) % IPU3_CSS_POOL_SIZE; in imgu_css_pool_last()
|
| /linux/drivers/net/team/ |
| H A D | team_mode_broadcast.c | 17 struct team_port *last = NULL; in bc_transmit() local 24 if (last) { in bc_transmit() 27 ret = !team_dev_queue_xmit(team, last, in bc_transmit() 33 last = cur; in bc_transmit() 36 if (last) { in bc_transmit() 37 ret = !team_dev_queue_xmit(team, last, skb); in bc_transmit()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_range_fence.c | 14 #define XE_RANGE_TREE_LAST(_node) ((_node)->last) 57 u64 start, u64 last, struct dma_fence *fence) in xe_range_fence_insert() argument 68 rfence->last = last; in xe_range_fence_insert() 135 u64 last) in xe_range_fence_tree_first() argument 137 return xe_range_fence_tree_iter_first(&tree->root, start, last); in xe_range_fence_tree_first() 149 xe_range_fence_tree_next(struct xe_range_fence *rfence, u64 start, u64 last) in xe_range_fence_tree_next() argument 151 return xe_range_fence_tree_iter_next(rfence, start, last); in xe_range_fence_tree_next()
|
| /linux/drivers/gpu/drm/nouveau/nvkm/subdev/bios/ |
| H A D | image.c | 54 image->last = pcir.last; in nvbios_imagen() 60 image->last = npde.last; in nvbios_imagen() 62 image->last = true; in nvbios_imagen() 76 if (image->last || !nvbios_imagen(bios, image)) { in nvbios_image()
|
| /linux/include/linux/ |
| H A D | vhost_iotlb.h | 11 u64 last; member 33 int vhost_iotlb_add_range_ctx(struct vhost_iotlb *iotlb, u64 start, u64 last, 35 int vhost_iotlb_add_range(struct vhost_iotlb *iotlb, u64 start, u64 last, 37 void vhost_iotlb_del_range(struct vhost_iotlb *iotlb, u64 start, u64 last); 46 vhost_iotlb_itree_first(struct vhost_iotlb *iotlb, u64 start, u64 last); 48 vhost_iotlb_itree_next(struct vhost_iotlb_map *map, u64 start, u64 last);
|