/linux/net/rds/ |
H A D | message.c | 51 void rds_message_addref(struct rds_message *rm) in rds_message_addref() argument 53 rdsdebug("addref rm %p ref %d\n", rm, refcount_read(&rm->m_refcount)); in rds_message_addref() 54 refcount_inc(&rm->m_refcount); in rds_message_addref() 130 static void rds_message_purge(struct rds_message *rm) in rds_message_purge() argument 135 if (unlikely(test_bit(RDS_MSG_PAGEVEC, &rm->m_flags))) in rds_message_purge() 138 spin_lock_irqsave(&rm->m_rs_lock, flags); in rds_message_purge() 139 if (rm->m_rs) { in rds_message_purge() 140 struct rds_sock *rs = rm->m_rs; in rds_message_purge() 142 if (rm->data.op_mmp_znotifier) { in rds_message_purge() 144 rds_rm_zerocopy_callback(rs, rm->data.op_mmp_znotifier); in rds_message_purge() [all …]
|
H A D | send.c | 67 struct rds_message *rm, *tmp; in rds_send_path_reset() local 71 rm = cp->cp_xmit_rm; in rds_send_path_reset() 77 rds_message_unmapped(rm); in rds_send_path_reset() 78 rds_message_put(rm); in rds_send_path_reset() 95 list_for_each_entry_safe(rm, tmp, &cp->cp_retrans, m_conn_item) { in rds_send_path_reset() 96 set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags); in rds_send_path_reset() 97 set_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags); in rds_send_path_reset() 139 struct rds_message *rm; in rds_send_xmit() local 201 rm = cp->cp_xmit_rm; in rds_send_xmit() 203 if (!rm) { in rds_send_xmit() [all …]
|
H A D | tcp_send.c | 70 int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm, in rds_tcp_xmit() argument 73 struct rds_conn_path *cp = rm->m_inc.i_conn_path; in rds_tcp_xmit() 86 rm->m_ack_seq = tc->t_last_sent_nxt + in rds_tcp_xmit() 88 be32_to_cpu(rm->m_inc.i_hdr.h_len) - 1; in rds_tcp_xmit() 90 set_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags); in rds_tcp_xmit() 91 tc->t_last_expected_una = rm->m_ack_seq + 1; in rds_tcp_xmit() 93 if (test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags)) in rds_tcp_xmit() 94 rm->m_inc.i_hdr.h_flags |= RDS_FLAG_RETRANSMITTED; in rds_tcp_xmit() 96 rdsdebug("rm %p tcp nxt %u ack_seq %llu\n", in rds_tcp_xmit() 97 rm, rds_tcp_write_seq(tc), in rds_tcp_xmit() [all …]
|
H A D | rdma.c | 611 int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm, in rds_cmsg_rdma_args() argument 616 struct rm_rdma_op *op = &rm->rdma; in rds_cmsg_rdma_args() 626 || rm->rdma.op_active) in rds_cmsg_rdma_args() 672 op->op_sg = rds_message_alloc_sgs(rm, nr_pages); in rds_cmsg_rdma_args() 811 int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm, in rds_cmsg_rdma_dest() argument 820 rm->m_rdma_cookie != 0) in rds_cmsg_rdma_dest() 823 memcpy(&rm->m_rdma_cookie, CMSG_DATA(cmsg), sizeof(rm->m_rdma_cookie)); in rds_cmsg_rdma_dest() 830 r_key = rds_rdma_cookie_key(rm->m_rdma_cookie); in rds_cmsg_rdma_dest() 843 rm->rdma.op_rdma_mr = mr; in rds_cmsg_rdma_dest() 851 * in rm->m_rdma_cookie. This causes it to be sent along to the peer [all …]
|
H A D | ib_send.c | 48 static void rds_ib_send_complete(struct rds_message *rm, in rds_ib_send_complete() argument 50 void (*complete)(struct rds_message *rm, int status)) in rds_ib_send_complete() argument 70 complete(rm, notify_status); in rds_ib_send_complete() 146 * Returns the rm for no good reason other than it is unobtainable 154 struct rds_message *rm = NULL; in rds_ib_send_unmap_op() local 160 rm = container_of(send->s_op, struct rds_message, data); in rds_ib_send_unmap_op() 167 rm = container_of(send->s_op, struct rds_message, rdma); in rds_ib_send_unmap_op() 174 rm = container_of(send->s_op, struct rds_message, atomic); in rds_ib_send_unmap_op() 187 return rm; in rds_ib_send_unmap_op() 245 struct rds_message *rm = NULL; in rds_ib_send_cqe_handler() local [all …]
|
H A D | loop.c | 75 static int rds_loop_xmit(struct rds_connection *conn, struct rds_message *rm, in rds_loop_xmit() argument 79 struct scatterlist *sgp = &rm->data.op_sg[sg]; in rds_loop_xmit() 81 be32_to_cpu(rm->m_inc.i_hdr.h_len); in rds_loop_xmit() 84 if (rm->m_inc.i_hdr.h_flags & RDS_FLAG_CONG_BITMAP) { in rds_loop_xmit() 92 rds_inc_init(&rm->m_inc, conn, &conn->c_laddr); in rds_loop_xmit() 94 rds_message_addref(rm); in rds_loop_xmit() 96 rds_recv_incoming(conn, &conn->c_laddr, &conn->c_faddr, &rm->m_inc, in rds_loop_xmit() 99 rds_send_drop_acked(conn, be64_to_cpu(rm->m_inc.i_hdr.h_sequence), in rds_loop_xmit() 102 rds_inc_put(&rm->m_inc); in rds_loop_xmit() 108 * See rds_loop_xmit(). Since our inc is embedded in the rm, we [all …]
|
/linux/arch/arm/net/ |
H A D | bpf_jit_32.h | 165 #define _AL3_R(op, rd, rn, rm) ((op ## _R) | (rd) << 12 | (rn) << 16 | (rm)) argument 171 #define ARM_ADD_R(rd, rn, rm) _AL3_R(ARM_INST_ADD, rd, rn, rm) argument 172 #define ARM_ADDS_R(rd, rn, rm) _AL3_R(ARM_INST_ADDS, rd, rn, rm) argument 175 #define ARM_ADC_R(rd, rn, rm) _AL3_R(ARM_INST_ADC, rd, rn, rm) argument 178 #define ARM_AND_R(rd, rn, rm) _AL3_R(ARM_INST_AND, rd, rn, rm) argument 179 #define ARM_ANDS_R(rd, rn, rm) _AL3_R(ARM_INST_ANDS, rd, rn, rm) argument 182 #define ARM_BIC_R(rd, rn, rm) _AL3_R(ARM_INST_BIC, rd, rn, rm) argument 186 #define ARM_BX(rm) (ARM_INST_BX | (rm)) argument 187 #define ARM_BLX_R(rm) (ARM_INST_BLX_R | (rm)) argument 189 #define ARM_CMP_R(rn, rm) _AL3_R(ARM_INST_CMP, 0, rn, rm) argument [all …]
|
/linux/drivers/macintosh/ |
H A D | rack-meter.c | 51 struct rackmeter *rm; member 95 static void rackmeter_setup_i2s(struct rackmeter *rm) in rackmeter_setup_i2s() argument 97 struct macio_chip *macio = rm->mdev->bus->chip; in rackmeter_setup_i2s() 106 pmac_call_feature(PMAC_FTR_SOUND_CHIP_ENABLE, rm->i2s, 0, 1); in rackmeter_setup_i2s() 121 out_le32(rm->i2s_regs + 0x10, 0x01fa0000); in rackmeter_setup_i2s() 122 (void)in_le32(rm->i2s_regs + 0x10); in rackmeter_setup_i2s() 132 static void rackmeter_set_default_pattern(struct rackmeter *rm) in rackmeter_set_default_pattern() argument 138 rm->ubuf[i] = (i & 1) * 255; in rackmeter_set_default_pattern() 140 rm->ubuf[i] = ((~i) & 1) * 255; in rackmeter_set_default_pattern() 144 static void rackmeter_do_pause(struct rackmeter *rm, int pause) in rackmeter_do_pause() argument [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ |
H A D | vmm.c | 30 if (vmm->rm.external) { in r535_mmu_vaspace_del() 33 ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.device.object, in r535_mmu_vaspace_del() 37 ctrl->hVASpace = vmm->rm.object.handle; in r535_mmu_vaspace_del() 39 WARN_ON(nvkm_gsp_rm_ctrl_wr(&vmm->rm.device.object, ctrl)); in r535_mmu_vaspace_del() 42 vmm->rm.external = false; in r535_mmu_vaspace_del() 45 nvkm_gsp_rm_free(&vmm->rm.object); in r535_mmu_vaspace_del() 46 nvkm_gsp_device_dtor(&vmm->rm.device); in r535_mmu_vaspace_del() 47 nvkm_gsp_client_dtor(&vmm->rm.client); in r535_mmu_vaspace_del() 49 nvkm_vmm_put(vmm, &vmm->rm.rsvd); in r535_mmu_vaspace_del() 59 &vmm->rm.client, &vmm->rm.device); in r535_mmu_vaspace_new() [all …]
|
/linux/tools/perf/arch/x86/util/ |
H A D | perf_regs.c | 200 * rm[0]: +/-NUM(REG) in arch_sdt_arg_parse_op() 201 * rm[1]: +/- in arch_sdt_arg_parse_op() 202 * rm[2]: NUM in arch_sdt_arg_parse_op() 203 * rm[3]: ( in arch_sdt_arg_parse_op() 204 * rm[4]: REG in arch_sdt_arg_parse_op() 205 * rm[5]: ) in arch_sdt_arg_parse_op() 207 regmatch_t rm[6]; in arch_sdt_arg_parse_op() local 226 regexec(&sdt_op_regex, old_op, 6, rm, 0) || in arch_sdt_arg_parse_op() 227 rm[4].rm_eo - rm[4].rm_so > SDT_REG_NAME_SIZE) { in arch_sdt_arg_parse_op() 244 if (rm[3].rm_so != rm[3].rm_eo) { in arch_sdt_arg_parse_op() [all …]
|
/linux/sound/soc/codecs/ |
H A D | tscs42xx.h | 114 #define RM(m, b) ((m)<<(b)) macro 133 #define RM_HPVOLL RM(FM_HPVOLL, FB_HPVOLL) 156 #define RM_HPVOLR RM(FM_HPVOLR, FB_HPVOLR) 179 #define RM_SPKVOLL RM(FM_SPKVOLL, FB_SPKVOLL) 204 #define RM_SPKVOLR RM(FM_SPKVOLR, FB_SPKVOLR) 229 #define RM_DACVOLL RM(FM_DACVOLL, FB_DACVOLL) 254 #define RM_DACVOLR RM(FM_DACVOLR, FB_DACVOLR) 279 #define RM_ADCVOLL RM(FM_ADCVOLL, FB_ADCVOLL) 304 #define RM_ADCVOLR RM(FM_ADCVOLR, FB_ADCVOLR) 337 RM(FM_INVOLL_INMUTEL, FB_INVOLL_INMUTEL) [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/vfn/ |
H A D | r535.c | 24 #include <rm/gpu.h> 37 const struct nvkm_rm_gpu *gpu = device->gsp->rm->gpu; in r535_vfn_new() 38 struct nvkm_vfn_func *rm; in r535_vfn_new() local 41 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_vfn_new() 44 rm->dtor = r535_vfn_dtor; in r535_vfn_new() 45 rm->intr = &tu102_vfn_intr; in r535_vfn_new() 46 rm->user.addr = 0x030000; in r535_vfn_new() 47 rm->user.size = 0x010000; in r535_vfn_new() 48 rm->user.base.minver = -1; in r535_vfn_new() 49 rm->user.base.maxver = -1; in r535_vfn_new() [all …]
|
/linux/tools/perf/arch/arm64/util/ |
H A D | perf_regs.c | 98 regmatch_t rm[5]; in arch_sdt_arg_parse_op() local 104 if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) { in arch_sdt_arg_parse_op() 107 new_len += (int)(rm[1].rm_eo - rm[1].rm_so); in arch_sdt_arg_parse_op() 114 (int)(rm[1].rm_eo - rm[1].rm_so), old_op + rm[1].rm_so); in arch_sdt_arg_parse_op() 115 } else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) { in arch_sdt_arg_parse_op() 120 if (rm[2].rm_so == -1) in arch_sdt_arg_parse_op() 123 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in arch_sdt_arg_parse_op() 129 if (rm[2].rm_so == -1) in arch_sdt_arg_parse_op() 133 (int)(rm[2].rm_eo - rm[2].rm_so), in arch_sdt_arg_parse_op() 134 old_op + rm[2].rm_so); in arch_sdt_arg_parse_op()
|
/linux/arch/mips/math-emu/ |
H A D | dp_mul.c | 16 u64 rm; in ieee754dp_mul() local 104 /* rm = xm * ym, re = xe+ye basically */ in ieee754dp_mul() 116 * Multiply 64 bits xm, ym to give high 64 bits rm with stickness. in ieee754dp_mul() 143 rm = hrm | (lrm != 0); in ieee754dp_mul() 148 if ((s64) rm < 0) { in ieee754dp_mul() 149 rm = (rm >> (64 - (DP_FBITS + 1 + 3))) | in ieee754dp_mul() 150 ((rm << (DP_FBITS + 1 + 3)) != 0); in ieee754dp_mul() 153 rm = (rm >> (64 - (DP_FBITS + 1 + 3 + 1))) | in ieee754dp_mul() 154 ((rm << (DP_FBITS + 1 + 3 + 1)) != 0); in ieee754dp_mul() 156 assert(rm & (DP_HIDDEN_BIT << 3)); in ieee754dp_mul() [all …]
|
H A D | sp_mul.c | 16 unsigned int rm; in ieee754sp_mul() local 104 /* rm = xm * ym, re = xe+ye basically */ in ieee754sp_mul() 116 * Multiply 32 bits xm, ym to give high 32 bits rm with stickness. in ieee754sp_mul() 138 rm = hrm | (lrm != 0); in ieee754sp_mul() 143 if ((int) rm < 0) { in ieee754sp_mul() 144 rm = (rm >> (32 - (SP_FBITS + 1 + 3))) | in ieee754sp_mul() 145 ((rm << (SP_FBITS + 1 + 3)) != 0); in ieee754sp_mul() 148 rm = (rm >> (32 - (SP_FBITS + 1 + 3 + 1))) | in ieee754sp_mul() 149 ((rm << (SP_FBITS + 1 + 3 + 1)) != 0); in ieee754sp_mul() 151 assert(rm & (SP_HIDDEN_BIT << 3)); in ieee754sp_mul() [all …]
|
H A D | sp_div.c | 14 unsigned int rm; in ieee754sp_div() local 115 rm = 0; in ieee754sp_div() 121 rm |= bm; in ieee754sp_div() 128 rm <<= 1; in ieee754sp_div() 130 rm |= 1; /* have remainder, set sticky */ in ieee754sp_div() 132 assert(rm); in ieee754sp_div() 134 /* normalise rm to rounding precision ? in ieee754sp_div() 136 while ((rm >> (SP_FBITS + 3)) == 0) { in ieee754sp_div() 137 rm <<= 1; in ieee754sp_div() 141 return ieee754sp_format(xs == ys ? 0 : 1, re, rm); in ieee754sp_div()
|
H A D | dp_div.c | 14 u64 rm; in ieee754dp_div() local 115 rm = 0; in ieee754dp_div() 121 rm |= bm; in ieee754dp_div() 128 rm <<= 1; in ieee754dp_div() 130 rm |= 1; /* have remainder, set sticky */ in ieee754dp_div() 132 assert(rm); in ieee754dp_div() 135 * Normalise rm to rounding precision ? in ieee754dp_div() 137 while ((rm >> (DP_FBITS + 3)) == 0) { in ieee754dp_div() 138 rm <<= 1; in ieee754dp_div() 142 return ieee754dp_format(xs == ys ? 0 : 1, re, rm); in ieee754dp_div()
|
H A D | dp_simple.c | 22 oldrm = ieee754_csr.rm; in ieee754dp_neg() 23 ieee754_csr.rm = FPU_CSR_RD; in ieee754dp_neg() 25 ieee754_csr.rm = oldrm; in ieee754dp_neg() 40 oldrm = ieee754_csr.rm; in ieee754dp_abs() 41 ieee754_csr.rm = FPU_CSR_RD; in ieee754dp_abs() 46 ieee754_csr.rm = oldrm; in ieee754dp_abs()
|
/linux/arch/x86/math-emu/ |
H A D | reg_divide.c | 28 int FPU_div(int flags, int rm, int control_w) in FPU_div() argument 37 deststnr = rm; in FPU_div() 46 a = (FPU_REG *) rm; in FPU_div() 49 a = &st(rm); in FPU_div() 51 taga = FPU_gettagi(rm); in FPU_div() 58 b = (FPU_REG *) rm; in FPU_div() 61 b = &st(rm); in FPU_div() 63 tagb = FPU_gettagi(rm); in FPU_div() 133 return real_2op_NaN((FPU_REG *) rm, flags & 0x0f, 0, in FPU_div() 141 return real_2op_NaN(st0_ptr, tag, rm, in FPU_div() [all …]
|
/linux/arch/arm64/crypto/ |
H A D | sm3-ce-core.S | 16 .macro sm3partw1, rd, rn, rm argument 17 .inst 0xce60c000 | .L\rd | (.L\rn << 5) | (.L\rm << 16) 20 .macro sm3partw2, rd, rn, rm argument 21 .inst 0xce60c400 | .L\rd | (.L\rn << 5) | (.L\rm << 16) 24 .macro sm3ss1, rd, rn, rm, ra 25 .inst 0xce400000 | .L\rd | (.L\rn << 5) | (.L\ra << 10) | (.L\rm << 16) 28 .macro sm3tt1a, rd, rn, rm, imm2 29 .inst 0xce408000 | .L\rd | (.L\rn << 5) | ((\imm2) << 12) | (.L\rm << 16) 32 .macro sm3tt1b, rd, rn, rm, imm2 33 .inst 0xce408400 | .L\rd | (.L\rn << 5) | ((\imm2) << 12) | (.L\rm << 16) [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ |
H A D | r535.c | 36 struct nvkm_devinit_func *rm; in r535_devinit_new() local 39 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_devinit_new() 42 rm->dtor = r535_devinit_dtor; in r535_devinit_new() 43 rm->post = hw->post; in r535_devinit_new() 44 rm->disable = hw->disable; in r535_devinit_new() 46 ret = nv50_devinit_new_(rm, device, type, inst, pdevinit); in r535_devinit_new() 48 kfree(rm); in r535_devinit_new()
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
H A D | r535.c | 71 struct nvkm_fb_func *rm; in r535_fb_new() local 74 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_fb_new() 77 rm->dtor = r535_fb_dtor; in r535_fb_new() 78 rm->sysmem.flush_page_init = hw->sysmem.flush_page_init; in r535_fb_new() 79 rm->vidmem.size = hw->vidmem.size; in r535_fb_new() 80 rm->ram_new = r535_fb_ram_new; in r535_fb_new() 82 ret = nvkm_fb_new_(rm, device, type, inst, pfb); in r535_fb_new() 84 kfree(rm); in r535_fb_new()
|
/linux/tools/virtio/ringtest/ |
H A D | Makefile | 23 -rm main.o 24 -rm ring.o ring 25 -rm virtio_ring_0_9.o virtio_ring_0_9 26 -rm virtio_ring_poll.o virtio_ring_poll 27 -rm virtio_ring_inorder.o virtio_ring_inorder 28 -rm ptr_ring.o ptr_ring 29 -rm noring.o noring
|
/linux/arch/sh/kernel/ |
H A D | traps_32.c | 107 unsigned long *rm, *rn; in handle_unaligned_ins() local 115 rm = ®s->regs[index]; in handle_unaligned_ins() 131 srcu = (unsigned char __user *)*rm; in handle_unaligned_ins() 145 src = (unsigned char *)rm; in handle_unaligned_ins() 158 case 1: /* mov.l Rm,@(disp,Rn) */ in handle_unaligned_ins() 159 src = (unsigned char*) rm; in handle_unaligned_ins() 171 src = (unsigned char*) rm; in handle_unaligned_ins() 181 case 5: /* mov.l @(disp,Rm),Rn */ in handle_unaligned_ins() 182 srcu = (unsigned char __user *)*rm; in handle_unaligned_ins() 193 srcu = (unsigned char __user *)*rm; in handle_unaligned_ins() [all …]
|
/linux/Documentation/devicetree/bindings/dma/ti/ |
H A D | k3-pktdma.yaml | 73 ti,sci-rm-range-tchan: 84 ti,sci-rm-range-tflow: 95 ti,sci-rm-range-rchan: 106 ti,sci-rm-range-rflow: 125 - ti,sci-rm-range-tchan 126 - ti,sci-rm-range-tflow 127 - ti,sci-rm-range-rchan 128 - ti,sci-rm-range-rflow 167 ti,sci-rm-range-tchan = <0x23>, /* UNMAPPED_TX_CHAN */ 173 ti,sci-rm-range-tflow = <0x10>, /* RING_UNMAPPED_TX_CHAN */ [all …]
|