| /linux/drivers/mailbox/ |
| H A D | bcm-pdc-mailbox.c | 385 struct pdc_rx_ctx rx_ctx[PDC_RING_ENTRIES]; member 587 struct pdc_rx_ctx *rx_ctx; in pdc_receive_one() local 600 (frags_rdy < pdcs->rx_ctx[pdcs->rxin].rxin_numd)) in pdc_receive_one() 616 rx_ctx = &pdcs->rx_ctx[rx_idx]; in pdc_receive_one() 617 num_frags = rx_ctx->rxin_numd; in pdc_receive_one() 619 mssg.ctx = rx_ctx->rxp_ctx; in pdc_receive_one() 620 rx_ctx->rxp_ctx = NULL; in pdc_receive_one() 621 resp_hdr = rx_ctx->resp_hdr; in pdc_receive_one() 622 resp_hdr_daddr = rx_ctx->resp_hdr_daddr; in pdc_receive_one() 623 dma_unmap_sg(dev, rx_ctx->dst_sg, sg_nents(rx_ctx->dst_sg), in pdc_receive_one() [all …]
|
| /linux/drivers/net/wireless/intel/ipw2x00/ |
| H A D | libipw_crypto_wep.c | 30 struct arc4_ctx rx_ctx; member 171 arc4_setkey(&wep->rx_ctx, key, klen); in libipw_wep_decrypt() 172 arc4_crypt(&wep->rx_ctx, pos, pos, plen + 4); in libipw_wep_decrypt()
|
| /linux/net/tls/ |
| H A D | tls_device.c | 699 struct tls_offload_context_rx *rx_ctx = tls_offload_ctx_rx(tls_ctx); in tls_device_resync_rx() local 702 trace_tls_device_rx_resync_send(sk, seq, rcd_sn, rx_ctx->resync_type); in tls_device_resync_rx() 773 struct tls_offload_context_rx *rx_ctx; in tls_device_rx_resync_new_rec() local 787 rx_ctx = tls_offload_ctx_rx(tls_ctx); in tls_device_rx_resync_new_rec() 790 switch (rx_ctx->resync_type) { in tls_device_rx_resync_new_rec() 792 resync_req = atomic64_read(&rx_ctx->resync_req); in tls_device_rx_resync_new_rec() 798 !atomic64_try_cmpxchg(&rx_ctx->resync_req, &resync_req, 0)) in tls_device_rx_resync_new_rec() 802 if (likely(!rx_ctx->resync_nh_do_now)) in tls_device_rx_resync_new_rec() 815 rx_ctx->resync_nh_do_now = 0; in tls_device_rx_resync_new_rec() 820 resync_req = atomic64_read(&rx_ctx->resync_async->req); in tls_device_rx_resync_new_rec() [all …]
|
| H A D | tls_sw.c | 1792 struct tls_sw_context_rx *rx_ctx = ctx->priv_ctx_rx; in tls_check_pending_rekey() local 1794 WRITE_ONCE(rx_ctx->key_update_pending, true); in tls_check_pending_rekey() 2680 struct tls_sw_context_rx *rx_ctx = tls_sw_ctx_rx(tls_ctx); in tls_sw_strparser_arm() local 2683 rx_ctx->saved_data_ready = sk->sk_data_ready; in tls_sw_strparser_arm() 2690 struct tls_sw_context_rx *rx_ctx = tls_sw_ctx_rx(tls_ctx); in tls_update_rx_zc_capable() local 2692 rx_ctx->zc_capable = tls_ctx->rx_no_pad || in tls_update_rx_zc_capable()
|
| H A D | tls_main.c | 770 struct tls_sw_context_rx *rx_ctx = tls_sw_ctx_rx(ctx); in do_tls_setsockopt_conf() local 772 tls_strp_check_rcv(&rx_ctx->strp); in do_tls_setsockopt_conf()
|
| /linux/include/net/ |
| H A D | tls.h | 450 struct tls_offload_context_rx *rx_ctx = tls_offload_ctx_rx(tls_ctx); in tls_offload_rx_resync_request() local 452 atomic64_set(&rx_ctx->resync_req, ((u64)ntohl(seq) << 32) | RESYNC_REQ); in tls_offload_rx_resync_request()
|
| /linux/net/tipc/ |
| H A D | crypto.c | 889 struct tipc_crypto_rx_ctx *rx_ctx; in tipc_aead_decrypt() local 911 ctx = tipc_aead_mem_alloc(tfm, sizeof(*rx_ctx), &iv, &req, &sg, nsg); in tipc_aead_decrypt() 943 rx_ctx = (struct tipc_crypto_rx_ctx *)ctx; in tipc_aead_decrypt() 944 rx_ctx->aead = aead; in tipc_aead_decrypt() 945 rx_ctx->bearer = b; in tipc_aead_decrypt() 969 struct tipc_crypto_rx_ctx *rx_ctx = TIPC_SKB_CB(skb)->crypto_ctx; in tipc_aead_decrypt_done() local 970 struct tipc_bearer *b = rx_ctx->bearer; in tipc_aead_decrypt_done() 971 struct tipc_aead *aead = rx_ctx->aead; in tipc_aead_decrypt_done() 986 kfree(rx_ctx); in tipc_aead_decrypt_done()
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_virtchnl_pf.c | 718 struct i40e_hmc_obj_rxq rx_ctx; in i40e_config_vsi_rx_queue() local 722 memset(&rx_ctx, 0, sizeof(struct i40e_hmc_obj_rxq)); in i40e_config_vsi_rx_queue() 725 rx_ctx.base = info->dma_ring_addr / 128; in i40e_config_vsi_rx_queue() 733 rx_ctx.qlen = info->ring_len; in i40e_config_vsi_rx_queue() 736 rx_ctx.hsplit_0 = I40E_RX_SPLIT_L2 | in i40e_config_vsi_rx_queue() 745 rx_ctx.hbuff = info->hdr_size >> I40E_RXQ_CTX_HBUFF_SHIFT; in i40e_config_vsi_rx_queue() 748 rx_ctx.dtype = I40E_RX_DTYPE_HEADER_SPLIT; in i40e_config_vsi_rx_queue() 756 rx_ctx.dbuff = info->databuffer_size >> I40E_RXQ_CTX_DBUFF_SHIFT; in i40e_config_vsi_rx_queue() 763 rx_ctx.rxmax = info->max_pkt_size; in i40e_config_vsi_rx_queue() 767 rx_ctx.rxmax += VLAN_HLEN; in i40e_config_vsi_rx_queue() [all …]
|
| H A D | i40e_main.c | 3571 struct i40e_hmc_obj_rxq rx_ctx; in i40e_configure_rx_ring() local 3579 memset(&rx_ctx, 0, sizeof(rx_ctx)); in i40e_configure_rx_ring() 3624 rx_ctx.dbuff = DIV_ROUND_UP(ring->rx_buf_len, in i40e_configure_rx_ring() 3627 rx_ctx.base = (ring->dma / 128); in i40e_configure_rx_ring() 3628 rx_ctx.qlen = ring->count; in i40e_configure_rx_ring() 3631 rx_ctx.dsize = 0; in i40e_configure_rx_ring() 3636 rx_ctx.hsplit_0 = 0; in i40e_configure_rx_ring() 3638 rx_ctx.rxmax = min_t(u16, vsi->max_frame, chain_len * ring->rx_buf_len); in i40e_configure_rx_ring() 3640 rx_ctx.lrxqthresh = 0; in i40e_configure_rx_ring() 3642 rx_ctx.lrxqthresh = 1; in i40e_configure_rx_ring() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
| H A D | ktls_rx.c | 433 struct tls_offload_context_rx *rx_ctx; in mlx5e_ktls_handle_get_psv_completion() local 440 rx_ctx = tls_offload_ctx_rx(tls_get_ctx(priv_rx->sk)); in mlx5e_ktls_handle_get_psv_completion() 441 async_resync = rx_ctx->resync_async; in mlx5e_ktls_handle_get_psv_completion()
|
| /linux/drivers/net/vmxnet3/ |
| H A D | vmxnet3_int.h | 334 struct vmxnet3_rx_ctx rx_ctx; member
|
| H A D | vmxnet3_drv.c | 1610 struct vmxnet3_rx_ctx *ctx = &rq->rx_ctx; in vmxnet3_rq_rx_complete() 2219 rq->rx_ctx.skb = NULL; in vmxnet3_rq_init()
|