| /linux/net/rxrpc/ |
| H A D | peer_event.c | 105 unsigned int max_data; in rxrpc_adjust_mtu() local 125 max_data = max_t(int, mtu - peer->hdrsize, 500); in rxrpc_adjust_mtu() 126 if (max_data < peer->max_data) { in rxrpc_adjust_mtu() 127 if (peer->pmtud_good > max_data) in rxrpc_adjust_mtu() 128 peer->pmtud_good = max_data; in rxrpc_adjust_mtu() 129 if (peer->pmtud_bad > max_data + 1) in rxrpc_adjust_mtu() 130 peer->pmtud_bad = max_data + 1; in rxrpc_adjust_mtu() 132 trace_rxrpc_pmtud_reduce(peer, 0, max_data, rxrpc_pmtud_reduce_icmp); in rxrpc_adjust_mtu() 133 peer->max_data = max_data; in rxrpc_adjust_mtu() 386 unsigned int max_data = peer->max_data; in rxrpc_input_probe_for_pmtud() local [all …]
|
| H A D | peer_object.c | 164 if (peer->max_data < peer->if_mtu - peer->hdrsize) { in rxrpc_assess_MTU_size() 167 peer->max_data = peer->if_mtu - peer->hdrsize; in rxrpc_assess_MTU_size() 210 peer->max_data = umin(RXRPC_JUMBO(1), peer->if_mtu - peer->hdrsize); in rxrpc_assess_MTU_size() 213 peer->pmtud_trial = umin(peer->max_data, peer->pmtud_bad - 1); in rxrpc_assess_MTU_size() 278 peer->max_data = peer->if_mtu - peer->hdrsize; in rxrpc_init_peer()
|
| H A D | input.c | 791 unsigned int max_data, capacity; in rxrpc_input_ack_trailer() local 810 if (max_mtu < peer->max_data) { in rxrpc_input_ack_trailer() 813 peer->max_data = max_mtu; in rxrpc_input_ack_trailer() 816 max_data = umin(max_mtu, peer->max_data); in rxrpc_input_ack_trailer() 817 capacity = max_data; in rxrpc_input_ack_trailer()
|
| H A D | output.c | 223 max_mtu = umax(call->peer->max_data, rxrpc_rx_mtu); in rxrpc_fill_out_ack() 679 len >= sizeof(struct rxrpc_wire_header) + call->peer->max_data) { in rxrpc_send_data_packet() 744 _leave(" = %d [%u]", ret, call->peer->max_data); in rxrpc_send_data_packet()
|
| H A D | conn_event.c | 163 max_mtu = umax(conn->peer->max_data, rxrpc_rx_mtu); in rxrpc_conn_retransmit_call()
|
| H A D | proc.c | 304 peer->max_data, in rxrpc_peer_seq_show()
|
| H A D | ar-internal.h | 407 unsigned int max_data; /* Maximum packet data capacity for this peer */ member
|
| /linux/net/sctp/ |
| H A D | chunk.c | 150 size_t len, first_len, max_data, remaining; in sctp_datamsg_from_user() local 174 max_data = asoc->frag_point; in sctp_datamsg_from_user() 175 if (unlikely(!max_data)) { in sctp_datamsg_from_user() 176 max_data = sctp_min_frag_point(sctp_sk(asoc->base.sk), in sctp_datamsg_from_user() 179 __func__, asoc, max_data); in sctp_datamsg_from_user() 191 max_data -= SCTP_PAD4(sizeof(struct sctp_auth_chunk) + in sctp_datamsg_from_user() 207 first_len = max_data; in sctp_datamsg_from_user() 218 msg_len > max_data) in sctp_datamsg_from_user() 246 len = max_data; in sctp_datamsg_from_user()
|
| /linux/drivers/crypto/intel/qat/qat_common/ |
| H A D | adf_pfvf_vf_proto.c | 114 u8 max_data; in adf_vf2pf_blkmsg_data_req() local 122 max_data = ADF_VF2PF_SMALL_BLOCK_BYTE_MAX; in adf_vf2pf_blkmsg_data_req() 128 max_data = ADF_VF2PF_MEDIUM_BLOCK_BYTE_MAX; in adf_vf2pf_blkmsg_data_req() 134 max_data = ADF_VF2PF_LARGE_BLOCK_BYTE_MAX; in adf_vf2pf_blkmsg_data_req() 141 if (*data > max_data) { in adf_vf2pf_blkmsg_data_req()
|
| /linux/fs/gfs2/ |
| H A D | file.c | 1236 unsigned int tmp, max_data = max_blocks - 3 * (sdp->sd_max_height - 1); in calc_max_reserv() local 1238 for (tmp = max_data; tmp > sdp->sd_diptrs;) { in calc_max_reserv() 1240 max_data -= tmp; in calc_max_reserv() 1243 *data_blocks = max_data; in calc_max_reserv() 1244 *ind_blocks = max_blocks - max_data; in calc_max_reserv() 1245 *len = ((loff_t)max_data - 3) << sdp->sd_sb.sb_bsize_shift; in calc_max_reserv()
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_txrx.c | 2176 unsigned int max_data = IAVF_MAX_DATA_PER_TXD_ALIGNED; in iavf_tx_map() local 2186 max_data += -dma & (IAVF_MAX_READ_REQ_SIZE - 1); in iavf_tx_map() 2192 max_data, td_tag); in iavf_tx_map() 2202 dma += max_data; in iavf_tx_map() 2203 size -= max_data; in iavf_tx_map() 2205 max_data = IAVF_MAX_DATA_PER_TXD_ALIGNED; in iavf_tx_map()
|
| /linux/include/trace/events/ |
| H A D | rxrpc.h | 2500 __field(unsigned short, max_data) 2509 __entry->max_data = conn->peer->max_data; 2518 __entry->max_data, 2550 unsigned int max_data, enum rxrpc_pmtud_reduce_trace reason), 2552 TP_ARGS(peer, serial, max_data, reason), 2557 __field(unsigned int, max_data) 2564 __entry->max_data = max_data; 2571 __entry->serial, __entry->max_data)
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_txrx.c | 1428 unsigned int max_data = ICE_MAX_DATA_PER_TXD_ALIGNED; in ice_tx_map() local 1438 max_data += -dma & (ICE_MAX_READ_REQ_SIZE - 1); in ice_tx_map() 1446 ice_build_ctob(td_cmd, td_offset, max_data, in ice_tx_map() 1457 dma += max_data; in ice_tx_map() 1458 size -= max_data; in ice_tx_map() 1460 max_data = ICE_MAX_DATA_PER_TXD_ALIGNED; in ice_tx_map()
|
| /linux/drivers/infiniband/hw/ionic/ |
| H A D | ionic_controlpath.c | 1707 int max_data) in ionic_qp_sq_init_cmb() argument 1781 int max_data, int sq_spec, struct ib_udata *udata) in ionic_qp_sq_init() argument 1815 if (max_data < 0) in ionic_qp_sq_init() 1818 if (max_data > ionic_v1_send_wqe_max_data(dev->lif_cfg.max_stride, in ionic_qp_sq_init() 1851 wqe_size = ionic_v1_send_wqe_min_size(max_sge, max_data, in ionic_qp_sq_init() 1860 wqe_size = ionic_v1_send_wqe_min_size(max_sge, max_data, in ionic_qp_sq_init() 1886 ionic_qp_sq_init_cmb(dev, qp, udata, max_data); in ionic_qp_sq_init()
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_txrx.c | 3596 unsigned int max_data = I40E_MAX_DATA_PER_TXD_ALIGNED; in i40e_tx_map() local 3606 max_data += -dma & (I40E_MAX_READ_REQ_SIZE - 1); in i40e_tx_map() 3612 max_data, td_tag); in i40e_tx_map() 3623 dma += max_data; in i40e_tx_map() 3624 size -= max_data; in i40e_tx_map() 3626 max_data = I40E_MAX_DATA_PER_TXD_ALIGNED; in i40e_tx_map()
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_txrx.c | 2666 unsigned int max_data = IDPF_TX_MAX_DESC_DATA_ALIGNED; in idpf_tx_splitq_map() local 2724 max_data += -dma & (IDPF_TX_MAX_READ_REQ_SIZE - 1); in idpf_tx_splitq_map() 2727 max_data); in idpf_tx_splitq_map() 2742 dma += max_data; in idpf_tx_splitq_map() 2743 size -= max_data; in idpf_tx_splitq_map() 2748 max_data = IDPF_TX_MAX_DESC_DATA_ALIGNED; in idpf_tx_splitq_map()
|
| /linux/kernel/trace/ |
| H A D | trace.c | 1560 struct trace_array_cpu *max_data = per_cpu_ptr(max_buf->data, cpu); in __update_max_tr() local 1565 max_data->saved_latency = tr->max_latency; in __update_max_tr() 1566 max_data->critical_start = data->critical_start; in __update_max_tr() 1567 max_data->critical_end = data->critical_end; in __update_max_tr() 1569 strscpy(max_data->comm, tsk->comm); in __update_max_tr() 1570 max_data->pid = tsk->pid; in __update_max_tr() 1576 max_data->uid = current_uid(); in __update_max_tr() 1578 max_data->uid = task_uid(tsk); in __update_max_tr() 1580 max_data->nice = tsk->static_prio - 20 - MAX_RT_PRIO; in __update_max_tr() 1581 max_data->policy = tsk->policy; in __update_max_tr() [all …]
|