| /linux/rust/syn/ |
| H A D | path.rs | 18 pub segments: Punctuated<PathSegment, Token![::]>, 29 segments: Punctuated::new(), in from() 31 path.segments.push_value(segment.into()); in from() 86 && self.segments.len() == 1 in get_ident() 87 && self.segments[0].arguments.is_none() in get_ident() 89 Some(&self.segments[0].ident) in get_ident() 101 self.segments.first().unwrap().ident.span(), in require_ident() 102 self.segments.last().unwrap().ident.span(), in require_ident() 334 && ty.path.segments.len() == 1 in parse() 335 && match &ty.path.segments[0].arguments { in parse() [all …]
|
| H A D | meta.rs | 381 let start_span = self.path.segments[0].ident.span(); in error() 408 segments: { in parse_meta_path() 409 let mut segments = Punctuated::new(); in parse_meta_path() localVariable 412 segments.push_value(PathSegment::from(ident)); in parse_meta_path() 422 segments.push_punct(punct); in parse_meta_path() 424 segments.push_value(PathSegment::from(ident)); in parse_meta_path() 426 segments in parse_meta_path()
|
| H A D | attr.rs | 250 path.segments.first().unwrap().ident.span(), in parse_args_with() 251 path.segments.last().unwrap().ident.span(), in parse_args_with() 538 path.segments.first().unwrap().ident.span(), in require_list() 539 path.segments.last().unwrap().ident.span(), in require_list() 556 path.segments.first().unwrap().ident.span(), in require_name_value() 557 path.segments.last().unwrap().ident.span(), in require_name_value() 779 for (i, segment) in self.0.segments.iter().enumerate() { in fmt()
|
| /linux/rust/macros/ |
| H A D | paste.rs | 7 let mut segments = Vec::new(); in concat_helper() localVariable 19 segments.push((value, lit.span())); in concat_helper() 26 segments.push((value, ident.span())); in concat_helper() 33 let (mut value, sp) = segments.pop().expect("expected identifier before modifier"); in concat_helper() 47 segments.push((value, sp)); in concat_helper() 51 segments.append(&mut concat_helper(tokens.as_slice())); in concat_helper() 53 token => panic!("unexpected token in paste segments: {token:?}"), in concat_helper() 57 segments in concat_helper() 61 let segments = concat_helper(tokens); in concat() localVariable 62 let pasted: String = segments in concat() [all...] |
| /linux/kernel/ |
| H A D | kexec.c | 24 struct kexec_segment *segments, in kimage_alloc_init() argument 47 memcpy(image->segment, segments, nr_segments * sizeof(*segments)); in kimage_alloc_init() 92 struct kexec_segment *segments, unsigned long flags) in do_kexec_load() argument 130 ret = kimage_alloc_init(&image, entry, nr_segments, segments, flags); in do_kexec_load() 243 struct kexec_segment __user *, segments, unsigned long, flags) in SYSCALL_DEFINE4() 257 ksegments = memdup_array_user(segments, nr_segments, sizeof(ksegments[0])); in SYSCALL_DEFINE4() 270 struct compat_kexec_segment __user *, segments, in COMPAT_SYSCALL_DEFINE4() argument 292 result = copy_from_user(&in, &segments[i], sizeof(in)); in COMPAT_SYSCALL_DEFINE4()
|
| /linux/Documentation/arch/powerpc/ |
| H A D | pci_iov_resource_on_powernv.rst | 95 * It is divided into 256 segments of equal size. A table in the chip 108 more segments. 120 has 256 segments; however, there is no table for mapping a segment 135 trick, to match to those giant segments. 144 - We cannot "group" segments in HW, so if a device ends up using more 153 PEs" that are used for the remaining M64 segments. 189 equally-sized segments. The finest granularity possible is a 256MB 190 window with 1MB segments. VF BARs that are 1MB or larger could be 196 BARs span several segments. 202 like the M32 window, but the segments can't be individually mapped to [all …]
|
| /linux/arch/arm/mm/ |
| H A D | proc-arm940.S | 117 mov r1, #(CACHE_DSEGMENTS - 1) << 4 @ 4 segments 123 bcs 1b @ segments 3 to 0 172 mov r1, #(CACHE_DSEGMENTS - 1) << 4 @ 4 segments 178 bcs 1b @ segments 7 to 0 195 mov r1, #(CACHE_DSEGMENTS - 1) << 4 @ 4 segments 201 bcs 1b @ segments 7 to 0 218 mov r1, #(CACHE_DSEGMENTS - 1) << 4 @ 4 segments 224 bcs 1b @ segments 7 to 0 241 mov r1, #(CACHE_DSEGMENTS - 1) << 4 @ 4 segments 252 bcs 1b @ segments 7 to 0
|
| /linux/include/uapi/linux/ |
| H A D | rpl.h | 42 } segments; member 45 #define rpl_segaddr segments.addr 46 #define rpl_segdata segments.data
|
| /linux/drivers/net/ethernet/sfc/siena/ |
| H A D | tx.c | 145 unsigned int segments; in __efx_siena_enqueue_skb() local 150 segments = skb_is_gso(skb) ? skb_shinfo(skb)->gso_segs : 0; in __efx_siena_enqueue_skb() 151 if (segments == 1) in __efx_siena_enqueue_skb() 152 segments = 0; /* Don't use TSO for a single segment. */ in __efx_siena_enqueue_skb() 158 if (segments) { in __efx_siena_enqueue_skb() 173 if (!data_mapped && (efx_siena_tx_map_data(tx_queue, skb, segments))) in __efx_siena_enqueue_skb()
|
| H A D | tx_common.c | 433 struct sk_buff *segments, *next; in efx_siena_tx_tso_fallback() local 435 segments = skb_gso_segment(skb, 0); in efx_siena_tx_tso_fallback() 436 if (IS_ERR(segments)) in efx_siena_tx_tso_fallback() 437 return PTR_ERR(segments); in efx_siena_tx_tso_fallback() 441 skb_list_walk_safe(segments, skb, next) { in efx_siena_tx_tso_fallback()
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-fs-nilfs2 | 183 What: /sys/fs/nilfs2/<device>/segments/segments_number 187 Show number of segments on a volume. 189 What: /sys/fs/nilfs2/<device>/segments/blocks_per_segment 195 What: /sys/fs/nilfs2/<device>/segments/clean_segments 199 Show count of clean segments. 201 What: /sys/fs/nilfs2/<device>/segments/dirty_segments 205 Show count of dirty segments. 207 What: /sys/fs/nilfs2/<device>/segments/README 211 Describe attributes of /sys/fs/nilfs2/<device>/segments
|
| H A D | sysfs-driver-jz4780-efuse | 6 split into segments. The driver supports read only. 7 The segments are:
|
| /linux/drivers/dma/xilinx/ |
| H A D | xilinx_dma.c | 375 struct list_head segments; member 651 seg = list_first_entry(&desc->segments, in xilinx_dma_get_metadata_ptr() 843 INIT_LIST_HEAD(&desc->segments); in xilinx_dma_alloc_tx_descriptor() 866 list_for_each_entry_safe(segment, next, &desc->segments, node) { in xilinx_dma_free_tx_descriptor() 872 &desc->segments, node) { in xilinx_dma_free_tx_descriptor() 878 &desc->segments, node) { in xilinx_dma_free_tx_descriptor() 884 &desc->segments, node) { in xilinx_dma_free_tx_descriptor() 994 list_for_each(entry, &desc->segments) { in xilinx_dma_get_residue() 1035 seg = list_first_entry(&desc->segments, struct xilinx_axidma_tx_segment, in xilinx_dma_get_residue_axidma_direct_s2mm() 1450 list_for_each_entry(segment, &desc->segments, node) { in xilinx_vdma_start_transfer() [all …]
|
| /linux/drivers/md/dm-pcache/ |
| H A D | cache.c | 121 pos->cache_seg = &cache->segments[latest.cache_seg_id]; in cache_pos_decode() 141 cache->segments = kvzalloc_objs(struct pcache_cache_segment, in cache_init() 143 if (!cache->segments) { in cache_init() 173 kvfree(cache->segments); in cache_init() 181 kvfree(cache->segments); in cache_exit() 201 cache->key_head.cache_seg = &cache->segments[0]; in cache_tail_init() 274 prev_cache_seg = &cache->segments[i]; in cache_segs_init()
|
| /linux/drivers/net/ovpn/ |
| H A D | io.c | 356 struct sk_buff *segments, *curr, *next; in ovpn_net_xmit() local 392 segments = skb_gso_segment(skb, 0); in ovpn_net_xmit() 393 if (IS_ERR(segments)) { in ovpn_net_xmit() 394 ret = PTR_ERR(segments); in ovpn_net_xmit() 401 skb = segments; in ovpn_net_xmit()
|
| /linux/Documentation/i2c/muxes/ |
| H A D | i2c-mux-gpio.rst | 10 i2c-mux-gpio is an i2c mux driver providing access to I2C bus segments 34 bus, the number of bus segments to create and the GPIO pins used 37 E.G. something like this for a MUX providing 4 bus segments
|
| /linux/drivers/net/ethernet/sfc/ |
| H A D | tx_common.c | 473 struct sk_buff *segments, *next; in efx_tx_tso_fallback() local 475 segments = skb_gso_segment(skb, 0); in efx_tx_tso_fallback() 476 if (IS_ERR(segments)) in efx_tx_tso_fallback() 477 return PTR_ERR(segments); in efx_tx_tso_fallback() 481 skb_list_walk_safe(segments, skb, next) { in efx_tx_tso_fallback()
|
| /linux/drivers/block/xen-blkback/ |
| H A D | blkback.c | 707 struct grant_page **pages = req->segments; in xen_blkbk_unmap_and_respond() 905 rc = xen_blkbk_map(pending_req->ring, pending_req->segments, in xen_blkbk_map_seg() 920 struct blkif_request_segment *segments = NULL; in xen_blkbk_parse_indirect() local 938 if (segments) in xen_blkbk_parse_indirect() 939 kunmap_atomic(segments); in xen_blkbk_parse_indirect() 940 segments = kmap_atomic(pages[n/SEGS_PER_INDIRECT_FRAME]->page); in xen_blkbk_parse_indirect() 944 pending_req->segments[n]->gref = segments[i].gref; in xen_blkbk_parse_indirect() 946 first_sect = READ_ONCE(segments[i].first_sect); in xen_blkbk_parse_indirect() 947 last_sect = READ_ONCE(segments[i].last_sect); in xen_blkbk_parse_indirect() 959 if (segments) in xen_blkbk_parse_indirect() [all …]
|
| /linux/drivers/media/dvb-frontends/ |
| H A D | mb86a20s.c | 1344 static u32 interpolate_value(u32 value, const struct linear_segments *segments, in interpolate_value() argument 1351 if (value >= segments[0].x) in interpolate_value() 1352 return segments[0].y; in interpolate_value() 1353 if (value < segments[len-1].x) in interpolate_value() 1354 return segments[len-1].y; in interpolate_value() 1358 if (value == segments[i].x) in interpolate_value() 1359 return segments[i].y; in interpolate_value() 1360 if (value > segments[i].x) in interpolate_value() 1365 dy = segments[i].y - segments[i - 1].y; in interpolate_value() 1366 dx = segments[i - 1].x - segments[i].x; in interpolate_value() [all …]
|
| /linux/drivers/scsi/mpi3mr/ |
| H A D | mpi3mr_fw.c | 523 struct segments *segments = op_reply_q->q_segments; in mpi3mr_get_reply_desc() local 527 segments[reply_ci / op_reply_q->segment_qd].segment; in mpi3mr_get_reply_desc() 1929 struct segments *segments; in mpi3mr_free_op_req_q_segments() local 1931 segments = mrioc->req_qinfo[q_idx].q_segments; in mpi3mr_free_op_req_q_segments() 1932 if (!segments) in mpi3mr_free_op_req_q_segments() 1949 if (!segments[j].segment) in mpi3mr_free_op_req_q_segments() 1952 size, segments[j].segment, segments[j].segment_dma); in mpi3mr_free_op_req_q_segments() 1953 segments[j].segment = NULL; in mpi3mr_free_op_req_q_segments() 1973 struct segments *segments; in mpi3mr_free_op_reply_q_segments() local 1975 segments = mrioc->op_reply_qinfo[q_idx].q_segments; in mpi3mr_free_op_reply_q_segments() [all …]
|
| /linux/Documentation/networking/ |
| H A D | tcp_ao.rst | 8 segments between trusted peers. It adds a new TCP header option with 49 |replayed TCP segments | |Extension (SNE) and | 131 of segments with TCP-AO but that do not match an MKT. The initial default 135 Alternately, the configuration can be changed to discard segments with 141 segments with TCP-AO are not discarded solely because they include 145 segments with unknown key signatures are discarded with warnings logged. 153 >> All TCP segments MUST be checked against the set of MKTs for matching 175 by TCP-AO when processing received TCP segments as discussed in the segment 179 segments are received out of order, and is considered a feature of TCP-AO, 217 that would allow accepting segments without a sign (which would be insecure). [all …]
|
| /linux/drivers/gpu/drm/arm/ |
| H A D | malidp_crtc.c | 95 } segments[MALIDP_COEFFTAB_NUM_COEFFS] = { variable 132 delta_in = segments[i].end - segments[i].start; in malidp_generate_gamma_table() 134 out_start = drm_color_lut_extract(lut[segments[i].start].green, in malidp_generate_gamma_table() 136 out_end = drm_color_lut_extract(lut[segments[i].end].green, 12); in malidp_generate_gamma_table()
|
| /linux/drivers/bus/mhi/host/ |
| H A D | boot.c | 373 int segments = DIV_ROUND_UP(alloc_size, seg_size) + 1; in mhi_alloc_bhie_table() local 383 img_info->mhi_buf = kzalloc_objs(*img_info->mhi_buf, segments); in mhi_alloc_bhie_table() 389 for (i = 0; i < segments; i++, mhi_buf++) { in mhi_alloc_bhie_table() 393 if (i == segments - 1) in mhi_alloc_bhie_table() 404 img_info->bhi_vec = img_info->mhi_buf[segments - 1].buf; in mhi_alloc_bhie_table() 405 img_info->entries = segments; in mhi_alloc_bhie_table()
|
| /linux/Documentation/filesystems/ |
| H A D | nilfs2.rst | 116 segments. This ioctl is used in lssu, 121 segments. This ioctl is used by 123 cleaning operation of segments and reduce 149 NILFS_IOCTL_SET_ALLOC_RANGE Define lower limit of segments in bytes and 150 upper limit of segments in bytes. This ioctl 191 A nilfs2 volume is equally divided into a number of segments except 240 3) Segment usage file (sufile) -- Stores allocation state of segments
|
| /linux/Documentation/input/devices/ |
| H A D | yealink.rst | 73 '8' : Generic 7 segment digit with individual addressable segments 75 Reduced capability 7 segment digit, when segments are hard wired together. 76 '1' : 2 segments digit only able to produce a 1. 101 map_seg7 Read/Write, the 7 segments char set, common for all
|