/linux/rust/macros/ |
H A D | quote.rs | 36 /// the given span. 42 ($span:expr => $($tt:tt)*) => {{ 47 let span = $span; localVariable 48 quote_spanned!(@proc tokens span $($tt)*); 52 (@proc $v:ident $span:ident) => {}; 53 (@proc $v:ident $span:ident #$id:ident $($tt:tt)*) => { 57 quote_spanned!(@proc $v $span $($tt)*); 59 (@proc $v:ident $span:ident #(#$id:ident)* $($tt:tt)*) => { 65 quote_spanned!(@proc $v $span $($tt)*); 67 (@proc $v:ident $span:ident ( $($inner:tt)* ) $($tt:tt)*) => { [all …]
|
H A D | paste.rs | 3 use proc_macro::{Delimiter, Group, Ident, Spacing, Span, TokenTree}; 5 fn concat(tokens: &[TokenTree], group_span: Span) -> TokenTree { in concat() 8 let mut span = None; in concat() localVariable 19 segments.push((value, lit.span())); in concat() 26 segments.push((value, ident.span())); in concat() 35 // Set the overall span of concatenated token as current span in concat() 36 "span" => { in concat() 38 span.is_none(), in concat() 39 "span modifier should only appear at most once" in concat() 41 span = Some(sp); in concat() [all …]
|
/linux/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | spectrum_span.c | 46 struct mlxsw_sp_span *span; member 77 return atomic_read(&mlxsw_sp->span->active_entries_count); in mlxsw_sp_span_occ_get() 83 struct mlxsw_sp_span *span; in mlxsw_sp_span_init() local 90 span = kzalloc(struct_size(span, entries, entries_count), GFP_KERNEL); in mlxsw_sp_span_init() 91 if (!span) in mlxsw_sp_span_init() 93 refcount_set(&span->policer_id_base_ref_count, 0); in mlxsw_sp_span_init() 94 span->entries_count = entries_count; in mlxsw_sp_span_init() 95 atomic_set(&span->active_entries_count, 0); in mlxsw_sp_span_init() 96 mutex_init(&span->analyzed_ports_lock); in mlxsw_sp_span_init() 97 INIT_LIST_HEAD(&span->analyzed_ports_list); in mlxsw_sp_span_init() [all …]
|
H A D | spectrum_span.h | 16 /* SPAN session identifiers that correspond to MLXSW_TRAP_ID_MIRROR_SESSION<i> 17 * trap identifiers. The session identifier is an attribute of the SPAN agent, 30 struct mlxsw_sp_port *dest_port; /* NULL for unoffloaded SPAN. */
|
/linux/drivers/net/ethernet/marvell/prestera/ |
H A D | prestera_span.c | 37 list_add_tail(&entry->list, &port->sw->span->entries); in prestera_span_entry_create() 49 prestera_span_entry_find_by_id(struct prestera_span *span, u8 span_id) in prestera_span_entry_find_by_id() argument 53 list_for_each_entry(entry, &span->entries, list) { in prestera_span_entry_find_by_id() 62 prestera_span_entry_find_by_port(struct prestera_span *span, in prestera_span_entry_find_by_port() argument 67 list_for_each_entry(entry, &span->entries, list) { in prestera_span_entry_find_by_port() 82 entry = prestera_span_entry_find_by_port(sw->span, port); in prestera_span_get() 108 entry = prestera_span_entry_find_by_id(sw->span, span_id); in prestera_span_put() 171 struct prestera_span *span; in prestera_span_init() local 173 span = kzalloc(sizeof(*span), GFP_KERNEL); in prestera_span_init() 174 if (!span) in prestera_span_init() [all …]
|
/linux/drivers/scsi/megaraid/ |
H A D | megaraid_sas_fp.c | 123 u16 MR_LdSpanArrayGet(u32 ld, u32 span, struct MR_DRV_RAID_MAP_ALL *map) in MR_LdSpanArrayGet() argument 125 return le16_to_cpu(map->raidMap.ldSpanMap[ld].spanBlock[span].span.arrayRef); in MR_LdSpanArrayGet() 148 static struct MR_LD_SPAN *MR_LdSpanPtrGet(u32 ld, u32 span, in MR_LdSpanPtrGet() argument 151 return &map->raidMap.ldSpanMap[ld].spanBlock[span].span; in MR_LdSpanPtrGet() 338 dev_err(&instance->pdev->dev, "megasas: span map %x, pDrvRaidMap->totalSize : %x\n", in MR_ValidateMapInfo() 379 u32 span, j; in MR_GetSpanBlock() local 381 for (span = 0; span < raid->spanDepth; span++, pSpanBlock++) { in MR_GetSpanBlock() 398 return span; in MR_GetSpanBlock() 408 * This routine calculates the Span block for given row using spanset. 418 * span - Span number [all …]
|
H A D | mbox_defs.h | 595 * adap_span_40ld_t - 40LD span 607 * adap_span_8ld_t - 8LD span 645 * @span : span 649 adap_span_40ld_t span[SPAN_DEPTH_8_SPANS]; member 656 * @span : span 662 adap_span_8ld_t span[SPAN_DEPTH_8_SPANS]; member 669 * @span : span 675 adap_span_8ld_t span[SPAN_DEPTH_4_SPANS]; member
|
/linux/drivers/iio/dac/ |
H A D | ltc2664.c | 92 /* span code of the channel */ 93 u8 span; member 136 int span, fs; in ltc2664_scale_get() local 138 span = chan->span; in ltc2664_scale_get() 139 if (span < 0) in ltc2664_scale_get() 140 return span; in ltc2664_scale_get() 142 fs = span_helper[span][1] - span_helper[span][0]; in ltc2664_scale_get() 150 int span, fs; in ltc2672_scale_get() local 152 span = chan->span - 1; in ltc2672_scale_get() 153 if (span < 0) in ltc2672_scale_get() [all …]
|
H A D | ltc2688.c | 139 int ret, reg, span; in ltc2688_span_get() local 145 span = FIELD_GET(LTC2688_CH_SPAN_MSK, reg); in ltc2688_span_get() 147 if (span >= LTC2688_SPAN_RANGE_MAX) in ltc2688_span_get() 150 return span; in ltc2688_span_get() 160 int span, fs; in ltc2688_scale_get() local 162 span = ltc2688_span_get(st, c); in ltc2688_scale_get() 163 if (span < 0) in ltc2688_scale_get() 164 return span; in ltc2688_scale_get() 166 fs = ltc2688_span_helper[span][1] - ltc2688_span_helper[span][0]; in ltc2688_scale_get() 177 int span; in ltc2688_offset_get() local [all …]
|
H A D | ad5791.c | 68 int (*get_lin_comp) (unsigned int span); 210 static int ad5791_get_lin_comp(unsigned int span) in ad5791_get_lin_comp() argument 212 if (span <= 10000) in ad5791_get_lin_comp() 214 else if (span <= 12000) in ad5791_get_lin_comp() 216 else if (span <= 16000) in ad5791_get_lin_comp() 218 else if (span <= 19000) in ad5791_get_lin_comp() 224 static int ad5780_get_lin_comp(unsigned int span) in ad5780_get_lin_comp() argument 226 if (span <= 10000) in ad5780_get_lin_comp()
|
/linux/drivers/platform/surface/aggregator/ |
H A D | ssh_parser.c | 18 * @src: The span of data over which the CRC should be computed. 21 * Computes the CRC of the provided data span (@src), compares it to the CRC 39 * @src: The data span to check the start of. 47 * sshp_find_syn() - Find SSH SYN bytes in the given data span. 48 * @src: The data span to search in. 49 * @rem: The span (output) indicating the remaining data, starting with SSH 52 * Search for SSH SYN bytes in the given source span. If found, set the @rem 53 * span to the remaining data, starting with the first SYN bytes and capped by 54 * the source span length, and return %true. This function does not copy any 58 * If no SSH SYN bytes could be found, set the @rem span to the zero-length [all …]
|
H A D | ssh_parser.h | 123 * sshp_buf_span_from() - Initialize a span from the given buffer and offset. 124 * @buf: The buffer to create the span from. 125 * @offset: The offset in the buffer at which the span should start. 126 * @span: The span to initialize (output). 128 * Initializes the provided span to point to the memory at the given offset in 129 * the buffer, with the length of the span being capped by the number of bytes 138 struct ssam_span *span) in sshp_buf_span_from() argument 140 span->ptr = buf->ptr + offset; in sshp_buf_span_from() 141 span->len = buf->len - offset; in sshp_buf_span_from()
|
/linux/fs/netfs/ |
H A D | iterator.c | 106 * Select the span of a bvec iterator we're going to use. Limit it by both maximum 107 * size and maximum number of segments. Returns the size of the span in bytes. 114 size_t len, span = 0, n = iter->count; in netfs_limit_bvec() local 133 span += len; in netfs_limit_bvec() 136 if (span >= max_size || nsegs >= max_segs) in netfs_limit_bvec() 142 return min(span, max_size); in netfs_limit_bvec() 146 * Select the span of an xarray iterator we're going to use. Limit it by both 149 * Returns the size of the span in bytes. 158 size_t span = 0, n = iter->count; in netfs_limit_xarray() local 181 span += len; in netfs_limit_xarray() [all …]
|
/linux/Documentation/scheduler/ |
H A D | sched-domains.rst | 10 Each scheduling domain spans a number of CPUs (stored in the ->span field). 11 A domain's span MUST be a superset of it child's span (this restriction could 12 be relaxed if the need arises), and a base domain for CPU i MUST span at least 13 i. The top domain for each CPU will generally span all CPUs in the system 16 explicitly set. A sched domain's span means "balance process load among these 22 domain's span. The group pointed to by the ->groups pointer MUST contain the CPU 58 The "base" domain will "span" the first level of the hierarchy. In the case 59 of SMT, you'll span all siblings of the physical CPU, with each group being 62 In SMP, the parent of the base domain will span all physical CPUs in the 64 of the SMP domain will span the entire machine, with each group having the
|
/linux/drivers/iommu/iommufd/ |
H A D | double_span.h | 12 * up and reported based on the tree that provides the interval. The first span 27 /* 0 = hole, 1 = used span[0], 2 = used span[1], -1 done iteration */ 46 #define interval_tree_for_each_double_span(span, itree1, itree2, first_index, \ argument 48 for (interval_tree_double_span_iter_first(span, itree1, itree2, \ 50 !interval_tree_double_span_iter_done(span); \ 51 interval_tree_double_span_iter_next(span))
|
H A D | pages.c | 939 struct interval_tree_double_span_iter span; member 961 struct interval_tree_double_span_iter span; in pfn_reader_unpin() local 966 interval_tree_for_each_double_span(&span, &pages->access_itree, in pfn_reader_unpin() 968 if (span.is_used) in pfn_reader_unpin() 971 batch_unpin(&pfns->batch, pages, span.start_hole - start, in pfn_reader_unpin() 972 span.last_hole - span.start_hole + 1); in pfn_reader_unpin() 976 /* Process a single span to load it from the proper storage */ 979 struct interval_tree_double_span_iter *span = &pfns->span; in pfn_reader_fill_span() local 985 WARN_ON(span->last_used < start_index)) in pfn_reader_fill_span() 988 if (span->is_used == 1) { in pfn_reader_fill_span() [all …]
|
H A D | io_pagetable.c | 73 static bool __alloc_iova_check_hole(struct interval_tree_double_span_iter *span, in __alloc_iova_check_hole() argument 78 if (span->is_used || span->last_hole - span->start_hole < length - 1) in __alloc_iova_check_hole() 81 span->start_hole = ALIGN(span->start_hole, iova_alignment) | in __alloc_iova_check_hole() 83 if (span->start_hole > span->last_hole || in __alloc_iova_check_hole() 84 span->last_hole - span->start_hole < length - 1) in __alloc_iova_check_hole() 89 static bool __alloc_iova_check_used(struct interval_tree_span_iter *span, in __alloc_iova_check_used() argument 94 if (span->is_hole || span->last_used - span->start_used < length - 1) in __alloc_iova_check_used() 97 span->start_used = ALIGN(span->start_used, iova_alignment) | in __alloc_iova_check_used() 99 if (span->start_used > span->last_used || in __alloc_iova_check_used() 100 span->last_used - span->start_used < length - 1) in __alloc_iova_check_used()
|
/linux/tools/testing/selftests/resctrl/ |
H A D | cmt_test.c | 77 static int check_results(struct resctrl_val_param *param, size_t span, int no_of_bits) in check_results() argument 108 return show_results_info(sum_llc_occu_resc, no_of_bits, span, in check_results() 126 size_t span; in cmt_run_test() local 156 span = cache_portion_size(cache_total_size, param.mask, long_mask); in cmt_run_test() 159 /* Duplicate the command to be able to replace span in it */ in cmt_run_test() 164 ret = asprintf(&span_str, "%zu", span); in cmt_run_test() 177 ret = check_results(¶m, span, n); in cmt_run_test()
|
H A D | mbm_test.c | 18 show_bw_info(unsigned long *bw_imc, unsigned long *bw_resc, size_t span) in show_bw_info() argument 43 ksft_print_msg("Span (MB): %zu\n", span / MB); in show_bw_info() 50 static int check_results(size_t span) in check_results() argument 82 ret = show_bw_info(bw_imc, bw_resc, span); in check_results()
|
H A D | cat_test.c | 141 * @span: Buffer size for the benchmark 159 size_t span, unsigned long current_mask) in cat_test() argument 192 buf = alloc_buffer(span, 1); in cat_test() 209 mem_flush(buf, span); in cat_test() 210 fill_cache_read(buf, span, true); in cat_test() 216 fill_cache_read(buf, span, true); in cat_test() 242 size_t span; in cat_run_test() local 278 span = cache_portion_size(cache_total_size, start_mask, full_cache_mask); in cat_run_test() 282 ret = cat_test(test, uparams, ¶m, span, start_mask); in cat_run_test()
|
/linux/include/linux/ |
H A D | interval_tree.h | 39 * nodes but classifies each span as either a hole, where no nodes intersect, or 82 #define interval_tree_for_each_span(span, itree, first_index, last_index) \ argument 83 for (interval_tree_span_iter_first(span, itree, \ 85 !interval_tree_span_iter_done(span); \ 86 interval_tree_span_iter_next(span))
|
/linux/lib/ |
H A D | interval_tree.c | 22 * span of nodes. This makes nodes[0]->last the end of that contiguous used span 24 * first node starting the next used span. A hole span is between nodes[0]->last 54 /* No nodes intersect the span, whole span is hole */ in interval_tree_span_iter_first()
|
/linux/tools/testing/selftests/bpf/progs/ |
H A D | test_access_variable_array.c | 8 unsigned long span = 0; variable 14 span = sd->span[0]; in BPF_PROG()
|
/linux/drivers/net/wireless/realtek/rtw89/ |
H A D | sar.c | 58 #define RTW89_SAR_SPAN_VALID(span) ((span)->subband_high) argument 70 /* Since 6GHz SAR subbands are not edge aligned, some cases span two SAR 92 const struct rtw89_sar_span *span = NULL; in rtw89_query_sar_config_common() local 100 * to make first span as index 0 of the table. So, if center in rtw89_query_sar_config_common() 104 span = &rtw89_sar_overlapping_6ghz[idx]; in rtw89_query_sar_config_common() 107 if (span && RTW89_SAR_SPAN_VALID(span)) { in rtw89_query_sar_config_common() 108 subband_l = span->subband_low; in rtw89_query_sar_config_common() 109 subband_h = span->subband_high; in rtw89_query_sar_config_common()
|
/linux/arch/sh/mm/ |
H A D | pmb.c | 146 unsigned long span; in pmb_mapping_exists() local 169 span = pmbe->size; in pmb_mapping_exists() 176 span += iter->size; in pmb_mapping_exists() 181 if (size <= span) { in pmb_mapping_exists() 381 * Link adjacent entries that span multiple PMB in pmb_bolt_mapping() 640 * see if the entries span a contiguous mapping. If so, in pmb_synchronize() 657 unsigned long span, newsize; in pmb_merge() local 661 span = newsize = head->size; in pmb_merge() 665 span += tail->size; in pmb_merge() 667 if (pmb_size_valid(span)) { in pmb_merge() [all …]
|