| /linux/drivers/net/wireless/mediatek/mt7601u/ |
| H A D | dma.c | 29 void *data, u32 seg_len, u32 truesize, struct page *p) in mt7601u_rx_skb_from_seg() argument 34 skb = alloc_skb(p ? 128 : seg_len, GFP_ATOMIC); in mt7601u_rx_skb_from_seg() 39 if (!true_len || true_len > seg_len) in mt7601u_rx_skb_from_seg() 77 u32 seg_len, struct page *p, in mt7601u_rx_process_seg() argument 82 u32 fce_info, truesize = seg_len; in mt7601u_rx_process_seg() 87 fce_info = get_unaligned_le32(data + seg_len - MT_FCE_INFO_LEN); in mt7601u_rx_process_seg() 88 seg_len -= MT_FCE_INFO_LEN; in mt7601u_rx_process_seg() 91 seg_len -= MT_DMA_HDR_LEN; in mt7601u_rx_process_seg() 95 seg_len -= sizeof(struct mt7601u_rxwi); in mt7601u_rx_process_seg() 104 skb = mt7601u_rx_skb_from_seg(dev, rxwi, data, seg_len, truesize, p); in mt7601u_rx_process_seg() [all …]
|
| /linux/drivers/net/ethernet/broadcom/bnxt/ |
| H A D | bnxt_coredump.c | 194 u16 segment_id, u32 *seg_len, in bnxt_hwrm_dbg_coredump_retrieve() argument 221 *seg_len = info.dest_buf_size; in bnxt_hwrm_dbg_coredump_retrieve() 229 struct coredump_segment_record *seg_rec, u32 seg_len, in bnxt_fill_coredump_seg_hdr() argument 246 seg_hdr->length = cpu_to_le32(seg_len); in bnxt_fill_coredump_seg_hdr() 369 size_t seg_len, extra_hlen = 0; in bnxt_get_ctx_coredump() local 387 seg_len = bnxt_copy_ctx_mem(bp, ctxm, data, 0) + extra_hlen; in bnxt_get_ctx_coredump() 389 bnxt_fill_coredump_seg_hdr(bp, &seg_hdr, NULL, seg_len, in bnxt_get_ctx_coredump() 395 buf += seg_len; in bnxt_get_ctx_coredump() 397 len += BNXT_SEG_HDR_LEN + seg_len; in bnxt_get_ctx_coredump() 461 u32 duration = 0, seg_len = 0; in __bnxt_get_coredump() local [all …]
|
| /linux/net/core/ |
| H A D | gso.c | 224 unsigned int seg_len, in skb_gso_size_check() argument 230 return seg_len <= max_len; in skb_gso_size_check() 233 seg_len -= GSO_BY_FRAGS; in skb_gso_size_check() 236 if (seg_len + skb_headlen(iter) > max_len) in skb_gso_size_check()
|
| H A D | skbuff.c | 1302 int i, len, seg_len; in skb_dump() local 1350 seg_len = min_t(int, skb_headlen(skb), len); in skb_dump() 1351 if (seg_len) in skb_dump() 1353 16, 1, skb->data, seg_len, false); in skb_dump() 1354 len -= seg_len; in skb_dump() 1377 seg_len = min_t(int, p_len, len); in skb_dump() 1381 16, 1, vaddr + p_off, seg_len, false); in skb_dump() 1383 len -= seg_len; in skb_dump()
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_hw_mbox.c | 373 u8 seq_id, u8 seg_len) in check_mbox_seq_id_and_seg_len() argument 375 if (seq_id > SEQ_ID_MAX_VAL || seg_len > MBOX_SEG_LEN) in check_mbox_seq_id_and_seg_len() 433 u8 seq_id, seg_len; in recv_mbox_handler() local 438 seg_len = HINIC_MBOX_HEADER_GET(mbox_header, SEG_LEN); in recv_mbox_handler() 441 if (!check_mbox_seq_id_and_seg_len(recv_mbox, seq_id, seg_len)) { in recv_mbox_handler() 444 src_func_idx, recv_mbox->seq_id, seq_id, seg_len); in recv_mbox_handler() 678 u16 seg_len) in mbox_copy_send_data() argument 686 if (seg_len % chk_sz) { in mbox_copy_send_data() 687 memcpy(mbox_max_buf, seg, seg_len); in mbox_copy_send_data() 691 data_len = seg_len; in mbox_copy_send_data() [all …]
|
| H A D | hinic_hw_mgmt.c | 492 int seq_id, seg_len; in recv_mgmt_msg_handler() local 496 seg_len = HINIC_MSG_HEADER_GET(*header, SEG_LEN); in recv_mgmt_msg_handler() 504 memcpy(recv_msg->msg + seq_id * SEGMENT_LEN, msg_body, seg_len); in recv_mgmt_msg_handler()
|
| /linux/fs/btrfs/ |
| H A D | lzo.c | 381 u32 seg_len; in lzo_decompress_bio() local 394 seg_len = read_compress_length(kaddr + offset_in_folio(cur_folio, cur_in)); in lzo_decompress_bio() 398 if (unlikely(seg_len > workspace_cbuf_length(fs_info))) { in lzo_decompress_bio() 408 cb->start, seg_len); in lzo_decompress_bio() 413 copy_compressed_segment(cb, workspace->cbuf, seg_len, &cur_in); in lzo_decompress_bio() 416 ret = lzo1x_decompress_safe(workspace->cbuf, seg_len, in lzo_decompress_bio()
|
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_desc.h | 32 __be16 seg_len; /* Length of this descriptor's segment */ member 49 __be16 seg_len; member
|
| /linux/lib/ |
| H A D | scatterlist.c | 437 unsigned int chunks, cur_page, seg_len, i, prv_len = 0; in sg_alloc_append_table_from_pages() local 479 seg_len = 0; in sg_alloc_append_table_from_pages() 481 seg_len += PAGE_SIZE; in sg_alloc_append_table_from_pages() 482 if (seg_len >= max_segment || in sg_alloc_append_table_from_pages() 485 seg_len = 0; in sg_alloc_append_table_from_pages() 495 seg_len = 0; in sg_alloc_append_table_from_pages() 497 seg_len += PAGE_SIZE; in sg_alloc_append_table_from_pages() 498 if (seg_len >= max_segment || in sg_alloc_append_table_from_pages()
|
| /linux/net/sched/ |
| H A D | sch_tbf.c | 211 unsigned int len = 0, prev_len = qdisc_pkt_len(skb), seg_len; in tbf_segment() local 222 seg_len = segs->len; in tbf_segment() 223 qdisc_skb_cb(segs)->pkt_len = seg_len; in tbf_segment() 231 len += seg_len; in tbf_segment()
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | svc_rdma_rw.c | 737 unsigned int sge_no, seg_len, len; in svc_rdma_build_read_segment() local 751 seg_len = min_t(unsigned int, len, in svc_rdma_build_read_segment() 758 seg_len, head->rc_pageoff); in svc_rdma_build_read_segment() 761 head->rc_pageoff += seg_len; in svc_rdma_build_read_segment() 766 len -= seg_len; in svc_rdma_build_read_segment()
|
| /linux/drivers/bus/mhi/host/ |
| H A D | boot.c | 372 size_t seg_size = mhi_cntrl->seg_len; in mhi_alloc_bhie_table() 449 if (mhi_cntrl->seg_len) in mhi_fw_load_type_get() 535 !mhi_cntrl->seg_len))) { in mhi_fw_load_handler()
|
| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| H A D | otx2_txrx.c | 932 int hdr_len, tcp_data, seg_len, pkt_len, offset; in otx2_sq_append_tso() local 954 seg_len = min_t(int, skb_shinfo(skb)->gso_size, tcp_data); in otx2_sq_append_tso() 955 tcp_data -= seg_len; in otx2_sq_append_tso() 965 tso_build_hdr(skb, hdr, &tso, seg_len, tcp_data == 0); in otx2_sq_append_tso() 973 while (seg_len > 0) { in otx2_sq_append_tso() 976 size = min_t(int, tso.size, seg_len); in otx2_sq_append_tso() 986 seg_len -= size; in otx2_sq_append_tso()
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | tid_rdma.c | 1093 u32 length = flow->req->seg_len; in kern_find_pages() 1117 flow->length = flow->req->seg_len - length; in kern_find_pages() 1749 req->cur_seg * req->seg_len + flow->sent); in hfi1_build_tid_rdma_read_packet() 1975 req->seg_len = qpriv->tid_rdma.local.max_len; in tid_rdma_rcv_read_request() 3107 rvt_skip_sge(&qpriv->tid_ss, (req->cur_seg * req->seg_len) + in hfi1_tid_rdma_restart_req() 3330 priv->tid_req.seg_len = in setup_tid_rdma_wqe() 3333 DIV_ROUND_UP(wqe->length, priv->tid_req.seg_len); in setup_tid_rdma_wqe() 3542 npkts = rvt_div_round_up_mtu(qp, req->seg_len); in hfi1_tid_write_alloc_resources() 3764 req->seg_len = qpriv->tid_rdma.local.max_len; in hfi1_rc_rcv_tid_rdma_write_req() 4119 flow->length = min_t(u32, req->seg_len, in hfi1_rc_rcv_tid_rdma_write_resp() [all …]
|
| H A D | tid_rdma.h | 113 u32 seg_len; member
|
| H A D | rc.c | 812 len = min_t(u32, req->seg_len, in hfi1_make_rc_req() 813 wqe->length - req->seg_len * req->cur_seg); in hfi1_make_rc_req() 1070 len = min_t(u32, req->seg_len, in hfi1_make_rc_req() 1071 wqe->length - req->seg_len * req->cur_seg); in hfi1_make_rc_req() 1116 len = min_t(u32, req->seg_len, in hfi1_make_rc_req() 1117 wqe->length - req->seg_len * req->cur_seg); in hfi1_make_rc_req()
|
| /linux/include/linux/qed/ |
| H A D | eth_common.h | 268 __le16 seg_len; member
|
| /linux/drivers/net/hippi/ |
| H A D | rrunner.h | 606 u32 seg_len; member
|
| /linux/fs/nfsd/ |
| H A D | nfs4proc.c | 3174 u32 seg_len = 2 * (1 + 2 + 1); in nfsd4_read_plus_rsize() local 3176 return (op_encode_hdr_size + 2 + seg_len + XDR_QUADLEN(rlen)) * sizeof(__be32); in nfsd4_read_plus_rsize()
|