| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_mw.c | 18 struct rxe_mw *mw = to_rmw(ibmw); in rxe_alloc_mw() local 25 ret = rxe_add_to_pool(&rxe->mw_pool, mw); in rxe_alloc_mw() 31 mw->rkey = ibmw->rkey = (mw->elem.index << 8) | rxe_get_next_key(-1); in rxe_alloc_mw() 32 mw->state = (mw->ibmw.type == IB_MW_TYPE_2) ? in rxe_alloc_mw() 34 spin_lock_init(&mw->lock); in rxe_alloc_mw() 36 rxe_finalize(mw); in rxe_alloc_mw() 43 struct rxe_mw *mw = to_rmw(ibmw); in rxe_dealloc_mw() local 45 rxe_cleanup(mw); in rxe_dealloc_mw() 51 struct rxe_mw *mw, struct rxe_mr *mr, int access) in rxe_check_bind_mw() argument 53 if (mw->ibmw.type == IB_MW_TYPE_1) { in rxe_check_bind_mw() [all …]
|
| /linux/net/netfilter/ipvs/ |
| H A D | ip_vs_wrr.c | 62 int mw; /* maximum weight */ member 118 mark->mw = ip_vs_wrr_max_weight(svc) - (mark->di - 1); in ip_vs_wrr_init_svc() 119 mark->cw = mark->mw; in ip_vs_wrr_init_svc() 145 mark->mw = ip_vs_wrr_max_weight(svc) - (mark->di - 1); in ip_vs_wrr_dest_changed() 146 if (mark->cw > mark->mw || !mark->cw) in ip_vs_wrr_dest_changed() 147 mark->cw = mark->mw; in ip_vs_wrr_dest_changed() 171 if (mark->mw == 0) in ip_vs_wrr_schedule() 187 mark->cw = mark->mw; in ip_vs_wrr_schedule()
|
| H A D | ip_vs_mh.c | 345 int mw, shift; in ip_vs_mh_shift_weight() local 363 mw = weight / gcd; in ip_vs_mh_shift_weight() 366 shift = fls(mw) - IP_VS_MH_TAB_BITS; in ip_vs_mh_shift_weight()
|
| /linux/arch/mips/txx9/generic/ |
| H A D | mem_tx4927.c | 46 unsigned int mw = 0; in tx4927_process_sdccr() local 61 mw = 8 >> sdccr_mw; in tx4927_process_sdccr() 64 return rs * cs * mw * bs; in tx4927_process_sdccr()
|
| /linux/include/uapi/rdma/ |
| H A D | rdma_user_rxe.h | 120 } mw; member
|
| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| H A D | otx2_pf.c | 299 void otx2_queue_vf_work(struct mbox *mw, struct workqueue_struct *mbox_wq, in otx2_queue_vf_work() argument 312 mbox = &mw->mbox; in otx2_queue_vf_work() 323 mw[i].num_msgs = hdr->num_msgs; in otx2_queue_vf_work() 325 queue_work(mbox_wq, &mw[i].mbox_wrk); in otx2_queue_vf_work() 328 mbox = &mw->mbox_up; in otx2_queue_vf_work() 332 mw[i].up_num_msgs = hdr->num_msgs; in otx2_queue_vf_work() 334 queue_work(mbox_wq, &mw[i].mbox_up_wrk); in otx2_queue_vf_work() 1047 struct mbox *mw = &pf->mbox; in otx2_pfaf_mbox_intr_handler() local 1063 mbox = &mw->mbox_up; in otx2_pfaf_mbox_intr_handler() 1069 queue_work(pf->mbox_wq, &mw->mbox_up_wrk); in otx2_pfaf_mbox_intr_handler() [all …]
|
| /linux/drivers/ntb/test/ |
| H A D | ntb_tool.c | 227 struct tool_mw *mw; member 869 struct tool_mw *outmw = outmw_wrap->mw; in tool_peer_mw_trans_read() 927 struct tool_mw *outmw = outmw_wrap->mw; in tool_peer_mw_trans_write() 999 tc->peers[pidx].outmws[widx].mw = &tc->outmws[widx]; in tool_init_mws()
|
| /linux/include/linux/mlx4/ |
| H A D | device.h | 1130 struct mlx4_mw *mw); 1131 void mlx4_mw_free(struct mlx4_dev *dev, struct mlx4_mw *mw); 1132 int mlx4_mw_enable(struct mlx4_dev *dev, struct mlx4_mw *mw);
|
| /linux/Documentation/driver-api/ |
| H A D | ntb.rst | 246 starts from the first mw idx, and a negative value starts from the last 247 mw idx. Both sides MUST set the same value here! The default value is
|
| /linux/drivers/infiniband/hw/ocrdma/ |
| H A D | ocrdma_stats.c | 161 (u64)rsrc_stats->mw); in ocrdma_resource_stats() 215 (u64)rsrc_stats->mw); in ocrdma_resource_stats()
|
| H A D | ocrdma_sli.h | 1951 u32 mw; member
|
| /linux/drivers/usb/typec/tcpm/ |
| H A D | tcpm.c | 4428 unsigned int mv, ma, mw, flags; in tcpm_pd_build_request() local 4459 mw = min_power(pdo, matching_snk_pdo); in tcpm_pd_build_request() 4460 ma = 1000 * mw / mv; in tcpm_pd_build_request() 4463 mw = ma * mv / 1000; in tcpm_pd_build_request() 4470 max_mw = mw; in tcpm_pd_build_request() 4471 if (mw < port->operating_snk_mw) { in tcpm_pd_build_request() 4487 *rdo = RDO_BATT(src_pdo_index + 1, mw, max_mw, flags); in tcpm_pd_build_request() 4490 src_pdo_index, mv, mw, in tcpm_pd_build_request()
|