/linux/drivers/md/ |
H A D | dm-writecache.c | 101 #define WC_MODE_PMEM(wc) ((wc)->pmem_mode) argument 102 #define WC_MODE_FUA(wc) ((wc)->writeback_fua) argument 104 #define WC_MODE_PMEM(wc) false argument 105 #define WC_MODE_FUA(wc) false argument 107 #define WC_MODE_SORT_FREELIST(wc) (!WC_MODE_PMEM(wc)) argument 225 struct dm_writecache *wc; member 234 struct dm_writecache *wc; member 243 static void wc_lock(struct dm_writecache *wc) in wc_lock() argument 245 mutex_lock(&wc->lock); in wc_lock() 248 static void wc_unlock(struct dm_writecache *wc) in wc_unlock() argument [all …]
|
/linux/include/math-emu/ |
H A D | op-common.h | 27 #define _FP_DECL(wc, X) \ argument 29 _FP_FRAC_DECL_##wc(X) 36 #define _FP_UNPACK_CANONICAL(fs, wc, X) \ argument 42 _FP_FRAC_SLL_##wc(X, _FP_WORKBITS); \ 48 if (_FP_FRAC_ZEROP_##wc(X)) \ 54 _FP_FRAC_CLZ_##wc(_shift, X); \ 56 _FP_FRAC_SLL_##wc(X, (_shift+_FP_WORKBITS)); \ 69 if (_FP_FRAC_ZEROP_##wc(X)) \ 89 #define _FP_PACK_CANONICAL(fs, wc, X) \ argument 97 _FP_ROUND(wc, X); \ [all …]
|
H A D | soft-fp.h | 135 #define _FP_ROUND_NEAREST(wc, X) \ argument 137 if ((_FP_FRAC_LOW_##wc(X) & 15) != _FP_WORK_ROUND) \ 138 _FP_FRAC_ADDI_##wc(X, _FP_WORK_ROUND); \ 141 #define _FP_ROUND_ZERO(wc, X) (void)0 argument 143 #define _FP_ROUND_PINF(wc, X) \ argument 145 if (!X##_s && (_FP_FRAC_LOW_##wc(X) & 7)) \ 146 _FP_FRAC_ADDI_##wc(X, _FP_WORK_LSB); \ 149 #define _FP_ROUND_MINF(wc, X) \ argument 151 if (X##_s && (_FP_FRAC_LOW_##wc(X) & 7)) \ 152 _FP_FRAC_ADDI_##wc(X, _FP_WORK_LSB); \ [all …]
|
/linux/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 512 struct ib_wc *wc) in mlx4_ib_handle_error_cqe() argument 526 wc->status = IB_WC_LOC_LEN_ERR; in mlx4_ib_handle_error_cqe() 529 wc->status = IB_WC_LOC_QP_OP_ERR; in mlx4_ib_handle_error_cqe() 532 wc->status = IB_WC_LOC_PROT_ERR; in mlx4_ib_handle_error_cqe() 535 wc->status = IB_WC_WR_FLUSH_ERR; in mlx4_ib_handle_error_cqe() 538 wc->status = IB_WC_MW_BIND_ERR; in mlx4_ib_handle_error_cqe() 541 wc->status = IB_WC_BAD_RESP_ERR; in mlx4_ib_handle_error_cqe() 544 wc->status = IB_WC_LOC_ACCESS_ERR; in mlx4_ib_handle_error_cqe() 547 wc->status = IB_WC_REM_INV_REQ_ERR; in mlx4_ib_handle_error_cqe() 550 wc->status = IB_WC_REM_ACCESS_ERR; in mlx4_ib_handle_error_cqe() [all …]
|
/linux/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 120 static void handle_good_req(struct ib_wc *wc, struct mlx5_cqe64 *cqe, in handle_good_req() argument 123 wc->wc_flags = 0; in handle_good_req() 126 wc->wc_flags |= IB_WC_WITH_IMM; in handle_good_req() 129 wc->opcode = IB_WC_RDMA_WRITE; in handle_good_req() 132 wc->wc_flags |= IB_WC_WITH_IMM; in handle_good_req() 136 wc->opcode = IB_WC_SEND; in handle_good_req() 139 wc->opcode = IB_WC_RDMA_READ; in handle_good_req() 140 wc->byte_len = be32_to_cpu(cqe->byte_cnt); in handle_good_req() 143 wc->opcode = IB_WC_COMP_SWAP; in handle_good_req() 144 wc->byte_len = 8; in handle_good_req() [all …]
|
H A D | gsi.c | 37 struct ib_wc wc; member 61 WARN_ON_ONCE(mlx5_ib_generate_wc(gsi_cq, &wr->wc)); in generate_completions() 68 static void handle_single_completion(struct ib_cq *cq, struct ib_wc *wc) in handle_single_completion() argument 72 container_of(wc->wr_cqe, struct mlx5_ib_gsi_wr, cqe); in handle_single_completion() 79 wr_id = wr->wc.wr_id; in handle_single_completion() 80 wr->wc = *wc; in handle_single_completion() 81 wr->wc.wr_id = wr_id; in handle_single_completion() 82 wr->wc.qp = &mqp->ibqp; in handle_single_completion() 360 struct ib_ud_wr *wr, struct ib_wc *wc) in mlx5_ib_add_outstanding_wr() argument 375 if (!wc) { in mlx5_ib_add_outstanding_wr() [all …]
|
/linux/fs/ocfs2/ |
H A D | aops.c | 787 static void ocfs2_unlock_pages(struct ocfs2_write_ctxt *wc) in ocfs2_unlock_pages() argument 796 if (wc->w_target_locked) { in ocfs2_unlock_pages() 797 BUG_ON(!wc->w_target_page); in ocfs2_unlock_pages() 798 for (i = 0; i < wc->w_num_pages; i++) { in ocfs2_unlock_pages() 799 if (wc->w_target_page == wc->w_pages[i]) { in ocfs2_unlock_pages() 800 wc->w_pages[i] = NULL; in ocfs2_unlock_pages() 804 mark_page_accessed(wc->w_target_page); in ocfs2_unlock_pages() 805 put_page(wc->w_target_page); in ocfs2_unlock_pages() 807 ocfs2_unlock_and_free_pages(wc->w_pages, wc->w_num_pages); in ocfs2_unlock_pages() 826 struct ocfs2_write_ctxt *wc) in ocfs2_free_write_ctxt() argument [all …]
|
/linux/drivers/net/ethernet/brocade/bna/ |
H A D | bfa_cs.h | 89 bfa_wc_up(struct bfa_wc *wc) in bfa_wc_up() argument 91 wc->wc_count++; in bfa_wc_up() 95 bfa_wc_down(struct bfa_wc *wc) in bfa_wc_down() argument 97 wc->wc_count--; in bfa_wc_down() 98 if (wc->wc_count == 0) in bfa_wc_down() 99 wc->wc_resume(wc->wc_cbarg); in bfa_wc_down() 104 bfa_wc_init(struct bfa_wc *wc, bfa_wc_resume_t wc_resume, void *wc_cbarg) in bfa_wc_init() argument 106 wc->wc_resume = wc_resume; in bfa_wc_init() 107 wc->wc_cbarg = wc_cbarg; in bfa_wc_init() 108 wc->wc_count = 0; in bfa_wc_init() [all …]
|
/linux/drivers/infiniband/sw/siw/ |
H A D | siw_cq.c | 48 int siw_reap_cqe(struct siw_cq *cq, struct ib_wc *wc) in siw_reap_cqe() argument 57 memset(wc, 0, sizeof(*wc)); in siw_reap_cqe() 58 wc->wr_id = cqe->id; in siw_reap_cqe() 59 wc->byte_len = cqe->bytes; in siw_reap_cqe() 68 wc->ex.invalidate_rkey = cqe->inval_stag; in siw_reap_cqe() 69 wc->wc_flags = IB_WC_WITH_INVALIDATE; in siw_reap_cqe() 71 wc->qp = cqe->base_qp; in siw_reap_cqe() 72 wc->opcode = map_wc_opcode[cqe->opcode]; in siw_reap_cqe() 73 wc->status = map_cqe_status[cqe->status].ib; in siw_reap_cqe() 95 wc->opcode = map_wc_opcode[opcode]; in siw_reap_cqe() [all …]
|
/linux/drivers/infiniband/hw/qib/ |
H A D | qib_ud.c | 62 struct ib_wc wc; in qib_ud_loopback() local 125 memset(&wc, 0, sizeof(wc)); in qib_ud_loopback() 126 wc.byte_len = length + sizeof(struct ib_grh); in qib_ud_loopback() 129 wc.wc_flags = IB_WC_WITH_IMM; in qib_ud_loopback() 130 wc.ex.imm_data = swqe->wr.ex.imm_data; in qib_ud_loopback() 155 if (unlikely(wc.byte_len > qp->r_len)) { in qib_ud_loopback() 168 wc.wc_flags |= IB_WC_GRH; in qib_ud_loopback() 201 wc.wr_id = qp->r_wr_id; in qib_ud_loopback() 202 wc.status = IB_WC_SUCCESS; in qib_ud_loopback() 203 wc.opcode = IB_WC_RECV; in qib_ud_loopback() [all …]
|
H A D | qib_uc.c | 246 struct ib_wc wc; in qib_uc_rcv() local 368 wc.ex.imm_data = ohdr->u.imm_data; in qib_uc_rcv() 370 wc.wc_flags = IB_WC_WITH_IMM; in qib_uc_rcv() 374 wc.ex.imm_data = 0; in qib_uc_rcv() 375 wc.wc_flags = 0; in qib_uc_rcv() 385 wc.byte_len = tlen + qp->r_rcv_len; in qib_uc_rcv() 386 if (unlikely(wc.byte_len > qp->r_len)) in qib_uc_rcv() 388 wc.opcode = IB_WC_RECV; in qib_uc_rcv() 392 wc.wr_id = qp->r_wr_id; in qib_uc_rcv() 393 wc.status = IB_WC_SUCCESS; in qib_uc_rcv() [all …]
|
/linux/drivers/infiniband/sw/rdmavt/ |
H A D | trace_cq.h | 70 TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), 71 TP_ARGS(cq, wc, idx), 85 __entry->wr_id = wc->wr_id; 86 __entry->status = wc->status; 87 __entry->opcode = wc->opcode; 88 __entry->length = wc->byte_len; 89 __entry->qpn = wc->qp->qp_num; 91 __entry->flags = wc->wc_flags; 92 __entry->imm = be32_to_cpu(wc->ex.imm_data); 110 TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), [all …]
|
/linux/tools/perf/tests/shell/ |
H A D | lock_contention.sh | 50 if [ "$(cat "${result}" | wc -l)" != "1" ]; then 51 echo "[Fail] Recorded result count is not 1:" "$(cat "${result}" | wc -l)" 68 if [ "$(cat "${result}" | wc -l)" != "1" ]; then 69 echo "[Fail] BPF result count is not 1:" "$(cat "${result}" | wc -l)" 80 if [ "$(cat "${result}" | wc -l)" != "1" ]; then 81 echo "[Fail] Recorded result count is not 1:" "$(cat "${result}" | wc -l)" 91 if [ "$(cat "${result}" | wc -l)" != "1" ]; then 92 echo "[Fail] Recorded result count is not 1:" "$(cat "${result}" | wc -l)" 103 if [ "$(cat "${result}" | wc -l)" != "1" ]; then 104 echo "[Fail] BPF result count is not 1:" "$(cat "${result}" | wc [all...] |
/linux/drivers/infiniband/hw/erdma/ |
H A D | erdma_cq.c | 110 static int erdma_poll_one_cqe(struct erdma_cq *cq, struct ib_wc *wc) in erdma_poll_one_cqe() argument 157 wc->wr_id = id_table[wqe_idx & (depth - 1)]; in erdma_poll_one_cqe() 158 wc->byte_len = be32_to_cpu(cqe->size); in erdma_poll_one_cqe() 160 wc->wc_flags = 0; in erdma_poll_one_cqe() 162 wc->opcode = wc_mapping_table[opcode]; in erdma_poll_one_cqe() 164 wc->ex.imm_data = cpu_to_be32(le32_to_cpu(cqe->imm_data)); in erdma_poll_one_cqe() 165 wc->wc_flags |= IB_WC_WITH_IMM; in erdma_poll_one_cqe() 167 wc->ex.invalidate_rkey = be32_to_cpu(cqe->inv_rkey); in erdma_poll_one_cqe() 168 wc->wc_flags |= IB_WC_WITH_INVALIDATE; in erdma_poll_one_cqe() 174 wc->status = map_cqe_status[syndrome].base; in erdma_poll_one_cqe() [all …]
|
/linux/Documentation/arch/x86/ |
H A D | pat.rst | 22 WC Write-combined 47 | ioremap_wc | -- | -- | WC | 54 | set_memory_wc, | WC | -- | -- | 62 | pci sysfs resource_wc | -- | -- | WC | 68 | pci proc | -- | -- | WC | 71 | /dev/mem | -- | WB/WC/UC- | WB/WC/UC- | 77 | /dev/mem | -- | WB/WC/UC- | WB/WC/UC- | 112 In addition, step 2 internally tracks the region as UC or WC in memtype 127 is made, should already have been ioremapped with WC attributes or PAT entries, 145 WC 000 WB _PAGE_CACHE_MODE_WB WC | WC [all …]
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | ud.c | 41 struct ib_wc wc; in ud_loopback() local 107 memset(&wc, 0, sizeof(wc)); in ud_loopback() 108 wc.byte_len = length + sizeof(struct ib_grh); in ud_loopback() 111 wc.wc_flags = IB_WC_WITH_IMM; in ud_loopback() 112 wc.ex.imm_data = swqe->wr.ex.imm_data; in ud_loopback() 137 if (unlikely(wc.byte_len > qp->r_len)) { in ud_loopback() 174 wc.wc_flags |= IB_WC_GRH; in ud_loopback() 193 wc.wr_id = qp->r_wr_id; in ud_loopback() 194 wc.status = IB_WC_SUCCESS; in ud_loopback() 195 wc.opcode = IB_WC_RECV; in ud_loopback() [all …]
|
H A D | uc.c | 270 struct ib_wc wc; in hfi1_uc_rcv() local 389 wc.ex.imm_data = ohdr->u.imm_data; in hfi1_uc_rcv() 390 wc.wc_flags = IB_WC_WITH_IMM; in hfi1_uc_rcv() 394 wc.ex.imm_data = 0; in hfi1_uc_rcv() 395 wc.wc_flags = 0; in hfi1_uc_rcv() 403 wc.byte_len = tlen + qp->r_rcv_len; in hfi1_uc_rcv() 404 if (unlikely(wc.byte_len > qp->r_len)) in hfi1_uc_rcv() 406 wc.opcode = IB_WC_RECV; in hfi1_uc_rcv() 410 wc.wr_id = qp->r_wr_id; in hfi1_uc_rcv() 411 wc.status = IB_WC_SUCCESS; in hfi1_uc_rcv() [all …]
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_cq.c | 324 struct ib_wc *wc) in pvrdma_poll_one() argument 357 wc->opcode = pvrdma_wc_opcode_to_ib(cqe->opcode); in pvrdma_poll_one() 358 wc->status = pvrdma_wc_status_to_ib(cqe->status); in pvrdma_poll_one() 359 wc->wr_id = cqe->wr_id; in pvrdma_poll_one() 360 wc->qp = &(*cur_qp)->ibqp; in pvrdma_poll_one() 361 wc->byte_len = cqe->byte_len; in pvrdma_poll_one() 362 wc->ex.imm_data = cqe->imm_data; in pvrdma_poll_one() 363 wc->src_qp = cqe->src_qp; in pvrdma_poll_one() 364 wc->wc_flags = pvrdma_wc_flags_to_ib(cqe->wc_flags); in pvrdma_poll_one() 365 wc->pkey_index = cqe->pkey_index; in pvrdma_poll_one() [all …]
|
/linux/drivers/infiniband/ulp/iser/ |
H A D | iser_initiator.c | 145 void (*done)(struct ib_cq *cq, struct ib_wc *wc)) in iser_create_send_desc() argument 533 void iser_login_rsp(struct ib_cq *cq, struct ib_wc *wc) in iser_login_rsp() argument 535 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_login_rsp() 537 struct iser_login_desc *desc = iser_login(wc->wr_cqe); in iser_login_rsp() 543 if (unlikely(wc->status != IB_WC_SUCCESS)) { in iser_login_rsp() 544 iser_err_comp(wc, "login_rsp"); in iser_login_rsp() 554 length = wc->byte_len - ISER_HEADERS_LEN; in iser_login_rsp() 592 static int iser_check_remote_inv(struct iser_conn *iser_conn, struct ib_wc *wc, in iser_check_remote_inv() argument 595 if (wc->wc_flags & IB_WC_WITH_INVALIDATE) { in iser_check_remote_inv() 597 u32 rkey = wc->ex.invalidate_rkey; in iser_check_remote_inv() [all …]
|
/linux/net/sunrpc/xprtrdma/ |
H A D | frwr_ops.c | 353 * frwr_wc_fastreg - Invoked by RDMA provider for a flushed FastReg WC 355 * @wc: WCE for a completed FastReg WR 359 static void frwr_wc_fastreg(struct ib_cq *cq, struct ib_wc *wc) in frwr_wc_fastreg() argument 361 struct ib_cqe *cqe = wc->wr_cqe; in frwr_wc_fastreg() 365 trace_xprtrdma_wc_fastreg(wc, &mr->mr_cid); in frwr_wc_fastreg() 367 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_fastreg() 442 static void frwr_mr_done(struct ib_wc *wc, struct rpcrdma_mr *mr) in frwr_mr_done() argument 444 if (likely(wc->status == IB_WC_SUCCESS)) in frwr_mr_done() 449 * frwr_wc_localinv - Invoked by RDMA provider for a LOCAL_INV WC 451 * @wc: WCE for a completed LocalInv WR [all …]
|
/linux/drivers/scsi/bfa/ |
H A D | bfa_cs.h | 242 bfa_wc_up(struct bfa_wc_s *wc) in bfa_wc_up() argument 244 wc->wc_count++; in bfa_wc_up() 248 bfa_wc_down(struct bfa_wc_s *wc) in bfa_wc_down() argument 250 wc->wc_count--; in bfa_wc_down() 251 if (wc->wc_count == 0) in bfa_wc_down() 252 wc->wc_resume(wc->wc_cbarg); in bfa_wc_down() 259 bfa_wc_init(struct bfa_wc_s *wc, bfa_wc_resume_t wc_resume, void *wc_cbarg) in bfa_wc_init() argument 261 wc->wc_resume = wc_resume; in bfa_wc_init() 262 wc->wc_cbarg = wc_cbarg; in bfa_wc_init() 263 wc->wc_count = 0; in bfa_wc_init() [all …]
|
/linux/fs/btrfs/ |
H A D | extent-tree.c | 5234 * @wc: the walk control for this deletion 5236 * @refs: the number of refs for wc->level - 1 5237 * @flags: the flags for wc->level - 1 5241 * wc->level should be read and walked into, or if we can simply delete our 5249 static bool visit_node_for_delete(struct btrfs_root *root, struct walk_control *wc, in visit_node_for_delete() argument 5254 int level = wc->level; in visit_node_for_delete() 5257 ASSERT(wc->refs[level - 1] > 0); in visit_node_for_delete() 5263 if (wc->stage == UPDATE_BACKREF) { in visit_node_for_delete() 5273 if (wc->refs[level - 1] == 1) in visit_node_for_delete() 5288 if (!wc->update_ref || generation <= btrfs_root_origin_generation(root)) in visit_node_for_delete() [all …]
|
/linux/drivers/infiniband/hw/cxgb4/ |
H A D | cq.c | 755 struct ib_wc *wc, struct c4iw_srq *srq) in __c4iw_poll_cq_one() argument 769 wc->wr_id = cookie; in __c4iw_poll_cq_one() 770 wc->qp = &qhp->ibqp; in __c4iw_poll_cq_one() 771 wc->vendor_err = CQE_STATUS(&cqe); in __c4iw_poll_cq_one() 772 wc->wc_flags = 0; in __c4iw_poll_cq_one() 790 wc->byte_len = CQE_LEN(&cqe); in __c4iw_poll_cq_one() 792 wc->byte_len = 0; in __c4iw_poll_cq_one() 796 wc->opcode = IB_WC_RECV; in __c4iw_poll_cq_one() 800 wc->opcode = IB_WC_RECV; in __c4iw_poll_cq_one() 801 wc->ex.invalidate_rkey = CQE_WRID_STAG(&cqe); in __c4iw_poll_cq_one() [all …]
|
/linux/net/smc/ |
H A D | smc_wr.c | 77 static inline void smc_wr_tx_process_cqe(struct ib_wc *wc) in smc_wr_tx_process_cqe() argument 83 link = wc->qp->qp_context; in smc_wr_tx_process_cqe() 85 if (wc->opcode == IB_WC_REG_MR) { in smc_wr_tx_process_cqe() 86 if (wc->status) in smc_wr_tx_process_cqe() 94 pnd_snd_idx = smc_wr_tx_find_pending_index(link, wc->wr_id); in smc_wr_tx_process_cqe() 97 link->wr_tx_v2_pend->wr_id != wc->wr_id) in smc_wr_tx_process_cqe() 99 link->wr_tx_v2_pend->wc_status = wc->status; in smc_wr_tx_process_cqe() 107 link->wr_tx_pends[pnd_snd_idx].wc_status = wc->status; in smc_wr_tx_process_cqe() 121 if (wc->status) { in smc_wr_tx_process_cqe() 132 pnd_snd.handler(&pnd_snd.priv, link, wc->status); in smc_wr_tx_process_cqe() [all …]
|
/linux/tools/testing/selftests/ftrace/test.d/filter/ |
H A D | event-filter-function.tc | 38 while [ `grep kmem_cache_free trace| wc -l` -eq 0 ]; do 46 hitcnt=`grep kmem_cache_free trace| grep $target_func | wc -l` 47 misscnt=`grep kmem_cache_free trace| grep -v $target_func | wc -l` 64 while [ `grep kmem_cache_free trace| wc -l` -eq 0 ]; do 72 hitcnt=`grep kmem_cache_free trace| grep $target_func | wc -l` 73 misscnt=`grep kmem_cache_free trace| grep -v $target_func | wc -l`
|