| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_mw.c | 18 struct rxe_mw *mw = to_rmw(ibmw); in rxe_alloc_mw() local 25 ret = rxe_add_to_pool(&rxe->mw_pool, mw); in rxe_alloc_mw() 31 mw->rkey = ibmw->rkey = (mw->elem.index << 8) | rxe_get_next_key(-1); in rxe_alloc_mw() 32 mw->state = (mw->ibmw.type == IB_MW_TYPE_2) ? in rxe_alloc_mw() 34 spin_lock_init(&mw->lock); in rxe_alloc_mw() 36 rxe_finalize(mw); in rxe_alloc_mw() 43 struct rxe_mw *mw = to_rmw(ibmw); in rxe_dealloc_mw() local 45 rxe_cleanup(mw); in rxe_dealloc_mw() 51 struct rxe_mw *mw, struct rxe_mr *mr, int access) in rxe_check_bind_mw() argument 53 if (mw->ibmw.type == IB_MW_TYPE_1) { in rxe_check_bind_mw() [all …]
|
| H A D | rxe_resp.c | 434 struct rxe_mw *mw = NULL; in check_rkey() local 486 mw = rxe_lookup_mw(qp, access, rkey); in check_rkey() 487 if (!mw) { in check_rkey() 493 mr = mw->mr; in check_rkey() 500 if (mw->access & IB_ZERO_BASED) in check_rkey() 501 qp->resp.offset = mw->addr; in check_rkey() 504 rxe_put(mw); in check_rkey() 505 mw = NULL; in check_rkey() 559 if (mw) in check_rkey() 560 rxe_put(mw); in check_rkey() [all …]
|
| /linux/net/netfilter/ipvs/ |
| H A D | ip_vs_wrr.c | 62 int mw; /* maximum weight */ member 118 mark->mw = ip_vs_wrr_max_weight(svc) - (mark->di - 1); in ip_vs_wrr_init_svc() 119 mark->cw = mark->mw; in ip_vs_wrr_init_svc() 145 mark->mw = ip_vs_wrr_max_weight(svc) - (mark->di - 1); in ip_vs_wrr_dest_changed() 146 if (mark->cw > mark->mw || !mark->cw) in ip_vs_wrr_dest_changed() 147 mark->cw = mark->mw; in ip_vs_wrr_dest_changed() 171 if (mark->mw == 0) in ip_vs_wrr_schedule() 187 mark->cw = mark->mw; in ip_vs_wrr_schedule()
|
| H A D | ip_vs_mh.c | 345 int mw, shift; in ip_vs_mh_shift_weight() local 363 mw = weight / gcd; in ip_vs_mh_shift_weight() 366 shift = fls(mw) - IP_VS_MH_TAB_BITS; in ip_vs_mh_shift_weight()
|
| /linux/drivers/ntb/ |
| H A D | ntb_transport.c | 585 struct ntb_transport_mw *mw; in ntb_transport_setup_qp_mw() local 597 mw = &nt->mw_vec[mw_num]; in ntb_transport_setup_qp_mw() 599 if (!mw->virt_addr) in ntb_transport_setup_qp_mw() 607 rx_size = (unsigned int)mw->xlat_size / num_qps_mw; in ntb_transport_setup_qp_mw() 608 qp->rx_buff = mw->virt_addr + rx_size * (qp_num / mw_count); in ntb_transport_setup_qp_mw() 761 struct ntb_transport_mw *mw = &nt->mw_vec[num_mw]; in ntb_free_mw() local 764 if (!mw->virt_addr) in ntb_free_mw() 768 dma_free_coherent(&pdev->dev, mw->alloc_size, in ntb_free_mw() 769 mw->alloc_addr, mw->dma_addr); in ntb_free_mw() 770 mw->xlat_size = 0; in ntb_free_mw() [all …]
|
| /linux/fs/ocfs2/ |
| H A D | dlmglue.c | 435 struct ocfs2_mask_waiter *mw, int ret) in ocfs2_update_lock_stats() argument 448 kt = ktime_sub(ktime_get(), mw->mw_lock_start); in ocfs2_update_lock_stats() 475 struct ocfs2_mask_waiter *mw; in ocfs2_track_lock_wait() local 482 mw = list_first_entry(&lockres->l_mask_waiters, in ocfs2_track_lock_wait() 485 ktime_to_us(ktime_mono_to_real(mw->mw_lock_start)); in ocfs2_track_lock_wait() 488 static inline void ocfs2_init_start_time(struct ocfs2_mask_waiter *mw) in ocfs2_init_start_time() argument 490 mw->mw_lock_start = ktime_get(); in ocfs2_init_start_time() 497 int level, struct ocfs2_mask_waiter *mw, int ret) in ocfs2_update_lock_stats() argument 506 static inline void ocfs2_init_start_time(struct ocfs2_mask_waiter *mw) in ocfs2_init_start_time() argument 892 struct ocfs2_mask_waiter *mw, *tmp; in lockres_set_flags() local [all …]
|
| /linux/arch/mips/txx9/generic/ |
| H A D | mem_tx4927.c | 46 unsigned int mw = 0; in tx4927_process_sdccr() local 61 mw = 8 >> sdccr_mw; in tx4927_process_sdccr() 64 return rs * cs * mw * bs; in tx4927_process_sdccr()
|
| /linux/drivers/net/ethernet/marvell/octeontx2/af/ |
| H A D | rvu.c | 35 static int rvu_mbox_init(struct rvu *rvu, struct mbox_wq_info *mw, 2246 struct mbox_wq_info *mw; in __rvu_mbox_handler() local 2251 mw = &rvu->afpf_wq_info; in __rvu_mbox_handler() 2254 mw = &rvu->afvf_wq_info; in __rvu_mbox_handler() 2260 devid = mwork - mw->mbox_wrk; in __rvu_mbox_handler() 2261 mbox = &mw->mbox; in __rvu_mbox_handler() 2266 if (mw->mbox_wrk[devid].num_msgs == 0) in __rvu_mbox_handler() 2272 req_hdr->opt_msg = mw->mbox_wrk[devid].num_msgs; in __rvu_mbox_handler() 2287 for (id = 0; id < mw->mbox_wrk[devid].num_msgs; id++) { in __rvu_mbox_handler() 2320 mw->mbox_wrk[devid].num_msgs = 0; in __rvu_mbox_handler() [all …]
|
| /linux/drivers/infiniband/hw/mlx4/ |
| H A D | mr.c | 358 struct mlx4_ib_mw *mw = to_mmw(ibmw); in mlx4_ib_alloc_mw() local 362 to_mlx4_type(ibmw->type), &mw->mmw); in mlx4_ib_alloc_mw() 366 err = mlx4_mw_enable(dev->dev, &mw->mmw); in mlx4_ib_alloc_mw() 370 ibmw->rkey = mw->mmw.key; in mlx4_ib_alloc_mw() 374 mlx4_mw_free(dev->dev, &mw->mmw); in mlx4_ib_alloc_mw() 380 struct mlx4_ib_mw *mw = to_mmw(ibmw); in mlx4_ib_dealloc_mw() local 382 mlx4_mw_free(to_mdev(ibmw->device)->dev, &mw->mmw); in mlx4_ib_dealloc_mw()
|
| /linux/drivers/clk/rockchip/ |
| H A D | clk.h | 796 #define COMPOSITE(_id, cname, pnames, f, mo, ms, mw, mf, ds, dw,\ argument 807 .mux_width = mw, \ 817 #define COMPOSITE_DIV_OFFSET(_id, cname, pnames, f, mo, ms, mw, \ argument 828 .mux_width = mw, \ 876 #define COMPOSITE_NODIV(_id, cname, pnames, f, mo, ms, mw, mf, \ argument 887 .mux_width = mw, \ 894 #define COMPOSITE_NOGATE(_id, cname, pnames, f, mo, ms, mw, mf, \ argument 905 .mux_width = mw, \ 914 mw, mf, ds, dw, df, dt) \ argument 924 .mux_width = mw, \ [all …]
|
| /linux/include/linux/usb/ |
| H A D | pd.h | 255 #define PDO_BATT_MAX_POWER(mw) ((((mw) / 250) & PDO_PWR_MASK) << PDO_BATT_MAX_PWR_SHIFT) argument 446 #define RDO_BATT_OP_PWR(mw) ((((mw) / 250) & RDO_PWR_MASK) << RDO_BATT_OP_PWR_SHIFT) argument 447 #define RDO_BATT_MAX_PWR(mw) ((((mw) / 250) & RDO_PWR_MASK) << RDO_BATT_MAX_PWR_SHIFT) argument
|
| /linux/include/dt-bindings/usb/ |
| H A D | pd.h | 44 #define PDO_BATT_MAX_POWER(mw) ((((mw) / 250) & PDO_PWR_MASK) << PDO_BATT_MAX_PWR_SHIFT) argument
|
| /linux/drivers/infiniband/hw/bnxt_re/ |
| H A D | ib_verbs.c | 509 fence->bind_rkey = ib_inc_rkey(fence->mw->rkey); in bnxt_re_create_fence_wqe() 550 if (fence->mw) { in bnxt_re_destroy_fence_mr() 551 bnxt_re_dealloc_mw(fence->mw); in bnxt_re_destroy_fence_mr() 552 fence->mw = NULL; in bnxt_re_destroy_fence_mr() 578 struct ib_mw *mw; in bnxt_re_create_fence_mr() local 630 mw = bnxt_re_alloc_mw(&pd->ib_pd, IB_MW_TYPE_1, NULL); in bnxt_re_create_fence_mr() 631 if (IS_ERR(mw)) { in bnxt_re_create_fence_mr() 634 rc = PTR_ERR(mw); in bnxt_re_create_fence_mr() 637 fence->mw = mw; in bnxt_re_create_fence_mr() 4182 struct bnxt_re_mw *mw; in bnxt_re_alloc_mw() local [all …]
|
| /linux/drivers/media/platform/nxp/dw100/ |
| H A D | dw100.c | 382 u32 sw, sh, mw, mh, idx; in dw100_ctrl_dewarping_map_init() local 389 mw = ctrl->dims[0]; in dw100_ctrl_dewarping_map_init() 394 qdx = qsw / (mw - 1); in dw100_ctrl_dewarping_map_init() 397 ctx->map_width = mw; in dw100_ctrl_dewarping_map_init() 399 ctx->map_size = mh * mw * sizeof(u32); in dw100_ctrl_dewarping_map_init() 402 qy = min_t(u32, (idx / mw) * qdy, qsh); in dw100_ctrl_dewarping_map_init() 403 qx = min_t(u32, (idx % mw) * qdx, qsw); in dw100_ctrl_dewarping_map_init()
|
| /linux/drivers/infiniband/core/ |
| H A D | uverbs_main.c | 118 int uverbs_dealloc_mw(struct ib_mw *mw) in uverbs_dealloc_mw() argument 120 struct ib_pd *pd = mw->pd; in uverbs_dealloc_mw() 123 ret = mw->device->ops.dealloc_mw(mw); in uverbs_dealloc_mw() 128 kfree(mw); in uverbs_dealloc_mw()
|
| H A D | uverbs_cmd.c | 926 struct ib_mw *mw; in ib_uverbs_alloc_mw() local 949 mw = rdma_zalloc_drv_obj(ib_dev, ib_mw); in ib_uverbs_alloc_mw() 950 if (!mw) { in ib_uverbs_alloc_mw() 955 mw->device = ib_dev; in ib_uverbs_alloc_mw() 956 mw->pd = pd; in ib_uverbs_alloc_mw() 957 mw->uobject = uobj; in ib_uverbs_alloc_mw() 958 mw->type = cmd.mw_type; in ib_uverbs_alloc_mw() 960 ret = pd->device->ops.alloc_mw(mw, &attrs->driver_udata); in ib_uverbs_alloc_mw() 966 uobj->object = mw; in ib_uverbs_alloc_mw() 970 resp.rkey = mw->rkey; in ib_uverbs_alloc_mw() [all …]
|
| /linux/include/uapi/rdma/ |
| H A D | rdma_user_rxe.h | 120 } mw; member
|
| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| H A D | otx2_pf.c | 299 void otx2_queue_vf_work(struct mbox *mw, struct workqueue_struct *mbox_wq, in otx2_queue_vf_work() argument 312 mbox = &mw->mbox; in otx2_queue_vf_work() 323 mw[i].num_msgs = hdr->num_msgs; in otx2_queue_vf_work() 325 queue_work(mbox_wq, &mw[i].mbox_wrk); in otx2_queue_vf_work() 328 mbox = &mw->mbox_up; in otx2_queue_vf_work() 332 mw[i].up_num_msgs = hdr->num_msgs; in otx2_queue_vf_work() 334 queue_work(mbox_wq, &mw[i].mbox_up_wrk); in otx2_queue_vf_work() 1047 struct mbox *mw = &pf->mbox; in otx2_pfaf_mbox_intr_handler() local 1063 mbox = &mw->mbox_up; in otx2_pfaf_mbox_intr_handler() 1069 queue_work(pf->mbox_wq, &mw->mbox_up_wrk); in otx2_pfaf_mbox_intr_handler() [all …]
|
| H A D | otx2_common.h | 453 void (*pf_queue_work_hdlr)(struct mbox *mb, struct workqueue_struct *mw, 1224 void otx2_queue_vf_work(struct mbox *mw, struct workqueue_struct *mbox_wq,
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | mr.c | 2542 struct mlx5_ib_mw *mw = to_mmw(ibmw); in mlx5_ib_alloc_mw() local 2582 err = mlx5_ib_create_mkey(dev, &mw->mmkey, in, inlen); in mlx5_ib_alloc_mw() 2586 mw->mmkey.type = MLX5_MKEY_MW; in mlx5_ib_alloc_mw() 2587 ibmw->rkey = mw->mmkey.key; in mlx5_ib_alloc_mw() 2588 mw->mmkey.ndescs = ndescs; in mlx5_ib_alloc_mw() 2599 err = mlx5r_store_odp_mkey(dev, &mw->mmkey); in mlx5_ib_alloc_mw() 2608 mlx5_core_destroy_mkey(dev->mdev, mw->mmkey.key); in mlx5_ib_alloc_mw() 2614 int mlx5_ib_dealloc_mw(struct ib_mw *mw) in mlx5_ib_dealloc_mw() argument 2616 struct mlx5_ib_dev *dev = to_mdev(mw->device); in mlx5_ib_dealloc_mw() 2617 struct mlx5_ib_mw *mmw = to_mmw(mw); in mlx5_ib_dealloc_mw()
|
| /linux/drivers/ntb/test/ |
| H A D | ntb_tool.c | 227 struct tool_mw *mw; member 869 struct tool_mw *outmw = outmw_wrap->mw; in tool_peer_mw_trans_read() 927 struct tool_mw *outmw = outmw_wrap->mw; in tool_peer_mw_trans_write() 999 tc->peers[pidx].outmws[widx].mw = &tc->outmws[widx]; in tool_init_mws()
|
| /linux/include/linux/mlx4/ |
| H A D | device.h | 1130 struct mlx4_mw *mw); 1131 void mlx4_mw_free(struct mlx4_dev *dev, struct mlx4_mw *mw); 1132 int mlx4_mw_enable(struct mlx4_dev *dev, struct mlx4_mw *mw);
|
| /linux/Documentation/driver-api/ |
| H A D | ntb.rst | 246 starts from the first mw idx, and a negative value starts from the last 247 mw idx. Both sides MUST set the same value here! The default value is
|
| /linux/drivers/infiniband/hw/ocrdma/ |
| H A D | ocrdma_stats.c | 161 (u64)rsrc_stats->mw); in ocrdma_resource_stats() 215 (u64)rsrc_stats->mw); in ocrdma_resource_stats()
|
| /linux/drivers/usb/typec/tcpm/ |
| H A D | tcpm.c | 4132 unsigned int mv, ma, mw, flags; in tcpm_pd_build_request() local 4163 mw = min_power(pdo, matching_snk_pdo); in tcpm_pd_build_request() 4164 ma = 1000 * mw / mv; in tcpm_pd_build_request() 4167 mw = ma * mv / 1000; in tcpm_pd_build_request() 4174 max_mw = mw; in tcpm_pd_build_request() 4175 if (mw < port->operating_snk_mw) { in tcpm_pd_build_request() 4191 *rdo = RDO_BATT(src_pdo_index + 1, mw, max_mw, flags); in tcpm_pd_build_request() 4194 src_pdo_index, mv, mw, in tcpm_pd_build_request()
|