/linux/drivers/gpu/drm/nouveau/include/nvrm/535.113.01/nvidia/kernel/inc/vgpu/ |
H A D | rpc_global_enums.h | 4 /* Excerpt of RM headers from https://github.com/NVIDIA/open-gpu-kernel-modules/tree/535.113.01 */ 11 X(RM, NOP) // 0 12 X(RM, SET_GUEST_SYSTEM_INFO) // 1 13 X(RM, ALLOC_ROOT) // 2 14 X(RM, ALLOC_DEVICE) // 3 deprecated 15 X(RM, ALLOC_MEMORY) // 4 16 X(RM, ALLOC_CTX_DMA) // 5 17 X(RM, ALLOC_CHANNEL_DMA) // 6 18 X(RM, MAP_MEMORY) // 7 19 X(RM, BIND_CTX_DMA) // 8 deprecated [all …]
|
/linux/net/rds/ |
H A D | message.c | 51 void rds_message_addref(struct rds_message *rm) in rds_message_addref() argument 53 rdsdebug("addref rm %p ref %d\n", rm, refcount_read(&rm->m_refcount)); in rds_message_addref() 54 refcount_inc(&rm->m_refcount); in rds_message_addref() 130 static void rds_message_purge(struct rds_message *rm) in rds_message_purge() argument 135 if (unlikely(test_bit(RDS_MSG_PAGEVEC, &rm->m_flags))) in rds_message_purge() 138 spin_lock_irqsave(&rm->m_rs_lock, flags); in rds_message_purge() 139 if (rm->m_rs) { in rds_message_purge() 140 struct rds_sock *rs = rm->m_rs; in rds_message_purge() 142 if (rm->data.op_mmp_znotifier) { in rds_message_purge() 144 rds_rm_zerocopy_callback(rs, rm->data.op_mmp_znotifier); in rds_message_purge() [all …]
|
H A D | send.c | 67 struct rds_message *rm, *tmp; in rds_send_path_reset() local 71 rm = cp->cp_xmit_rm; in rds_send_path_reset() 77 rds_message_unmapped(rm); in rds_send_path_reset() 78 rds_message_put(rm); in rds_send_path_reset() 95 list_for_each_entry_safe(rm, tmp, &cp->cp_retrans, m_conn_item) { in rds_send_path_reset() 96 set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags); in rds_send_path_reset() 97 set_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags); in rds_send_path_reset() 139 struct rds_message *rm; in rds_send_xmit() local 201 rm = cp->cp_xmit_rm; in rds_send_xmit() 203 if (!rm) { in rds_send_xmit() [all …]
|
H A D | tcp_send.c | 70 int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm, in rds_tcp_xmit() argument 73 struct rds_conn_path *cp = rm->m_inc.i_conn_path; in rds_tcp_xmit() 86 rm->m_ack_seq = tc->t_last_sent_nxt + in rds_tcp_xmit() 88 be32_to_cpu(rm->m_inc.i_hdr.h_len) - 1; in rds_tcp_xmit() 90 set_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags); in rds_tcp_xmit() 91 tc->t_last_expected_una = rm->m_ack_seq + 1; in rds_tcp_xmit() 93 if (test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags)) in rds_tcp_xmit() 94 rm->m_inc.i_hdr.h_flags |= RDS_FLAG_RETRANSMITTED; in rds_tcp_xmit() 96 rdsdebug("rm %p tcp nxt %u ack_seq %llu\n", in rds_tcp_xmit() 97 rm, rds_tcp_write_seq(tc), in rds_tcp_xmit() [all …]
|
H A D | rdma.c | 611 int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm, in rds_cmsg_rdma_args() argument 616 struct rm_rdma_op *op = &rm->rdma; in rds_cmsg_rdma_args() 626 || rm->rdma.op_active) in rds_cmsg_rdma_args() 672 op->op_sg = rds_message_alloc_sgs(rm, nr_pages); in rds_cmsg_rdma_args() 811 int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm, in rds_cmsg_rdma_dest() argument 820 rm->m_rdma_cookie != 0) in rds_cmsg_rdma_dest() 823 memcpy(&rm->m_rdma_cookie, CMSG_DATA(cmsg), sizeof(rm->m_rdma_cookie)); in rds_cmsg_rdma_dest() 830 r_key = rds_rdma_cookie_key(rm->m_rdma_cookie); in rds_cmsg_rdma_dest() 843 rm->rdma.op_rdma_mr = mr; in rds_cmsg_rdma_dest() 851 * in rm->m_rdma_cookie. This causes it to be sent along to the peer [all …]
|
H A D | ib_send.c | 48 static void rds_ib_send_complete(struct rds_message *rm, in rds_ib_send_complete() argument 50 void (*complete)(struct rds_message *rm, int status)) in rds_ib_send_complete() argument 70 complete(rm, notify_status); in rds_ib_send_complete() 146 * Returns the rm for no good reason other than it is unobtainable 154 struct rds_message *rm = NULL; in rds_ib_send_unmap_op() local 160 rm = container_of(send->s_op, struct rds_message, data); in rds_ib_send_unmap_op() 167 rm = container_of(send->s_op, struct rds_message, rdma); in rds_ib_send_unmap_op() 174 rm = container_of(send->s_op, struct rds_message, atomic); in rds_ib_send_unmap_op() 187 return rm; in rds_ib_send_unmap_op() 245 struct rds_message *rm = NULL; in rds_ib_send_cqe_handler() local [all …]
|
/linux/arch/arm/net/ |
H A D | bpf_jit_32.h | 165 #define _AL3_R(op, rd, rn, rm) ((op ## _R) | (rd) << 12 | (rn) << 16 | (rm)) argument 171 #define ARM_ADD_R(rd, rn, rm) _AL3_R(ARM_INST_ADD, rd, rn, rm) argument 172 #define ARM_ADDS_R(rd, rn, rm) _AL3_R(ARM_INST_ADDS, rd, rn, rm) argument 175 #define ARM_ADC_R(rd, rn, rm) _AL3_R(ARM_INST_ADC, rd, rn, rm) argument 178 #define ARM_AND_R(rd, rn, rm) _AL3_R(ARM_INST_AND, rd, rn, rm) argument 179 #define ARM_ANDS_R(rd, rn, rm) _AL3_R(ARM_INST_ANDS, rd, rn, rm) argument 182 #define ARM_BIC_R(rd, rn, rm) _AL3_R(ARM_INST_BIC, rd, rn, rm) argument 186 #define ARM_BX(rm) (ARM_INST_BX | (rm)) argument 187 #define ARM_BLX_R(rm) (ARM_INST_BLX_R | (rm)) argument 189 #define ARM_CMP_R(rn, rm) _AL3_R(ARM_INST_CMP, 0, rn, rm) argument [all …]
|
/linux/drivers/macintosh/ |
H A D | rack-meter.c | 51 struct rackmeter *rm; member 95 static void rackmeter_setup_i2s(struct rackmeter *rm) in rackmeter_setup_i2s() argument 97 struct macio_chip *macio = rm->mdev->bus->chip; in rackmeter_setup_i2s() 106 pmac_call_feature(PMAC_FTR_SOUND_CHIP_ENABLE, rm->i2s, 0, 1); in rackmeter_setup_i2s() 121 out_le32(rm->i2s_regs + 0x10, 0x01fa0000); in rackmeter_setup_i2s() 122 (void)in_le32(rm->i2s_regs + 0x10); in rackmeter_setup_i2s() 132 static void rackmeter_set_default_pattern(struct rackmeter *rm) in rackmeter_set_default_pattern() argument 138 rm->ubuf[i] = (i & 1) * 255; in rackmeter_set_default_pattern() 140 rm->ubuf[i] = ((~i) & 1) * 255; in rackmeter_set_default_pattern() 144 static void rackmeter_do_pause(struct rackmeter *rm, int pause) in rackmeter_do_pause() argument [all …]
|
/linux/drivers/gpu/drm/msm/disp/dpu1/ |
H A D | dpu_rm.c | 38 struct dpu_rm *rm, in dpu_rm_init() argument 45 if (!rm || !cat || !mmio) { in dpu_rm_init() 51 memset(rm, 0, sizeof(*rm)); in dpu_rm_init() 64 rm->mixer_blks[lm->id - LM_0] = &hw->base; in dpu_rm_init() 78 rm->merge_3d_blks[merge_3d->id - MERGE_3D_0] = &hw->base; in dpu_rm_init() 93 hw->merge_3d = to_dpu_hw_merge_3d(rm->merge_3d_blks[pp->merge_3d - MERGE_3D_0]); in dpu_rm_init() 94 rm->pingpong_blks[pp->id - PINGPONG_0] = &hw->base; in dpu_rm_init() 107 rm->hw_intf[intf->id - INTF_0] = hw; in dpu_rm_init() 120 rm->hw_wb[wb->id - WB_0] = hw; in dpu_rm_init() 133 rm->ctl_blks[ctl->id - CTL_0] = &hw->base; in dpu_rm_init() [all …]
|
H A D | dpu_rm.h | 44 * @rm: DPU Resource Manager handle 51 struct dpu_rm *rm, 63 * @rm: DPU Resource Manager handle 69 int dpu_rm_reserve(struct dpu_rm *rm, 78 * @rm: DPU Resource Manager handle 88 int dpu_rm_get_assigned_resources(struct dpu_rm *rm, 93 * dpu_rm_print_state - output the RM private state 102 * @rm: DPU Resource Manager handle 105 static inline struct dpu_hw_intf *dpu_rm_get_intf(struct dpu_rm *rm, enum dpu_intf intf_idx) in dpu_rm_get_intf() argument 107 return rm->hw_intf[intf_idx - INTF_0]; in dpu_rm_get_intf() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
H A D | r535.c | 77 nvkm_gsp_rm_free(&chan->rm.object); in r535_chan_fini() 131 args = nvkm_gsp_rm_alloc_get(&chan->disp->rm.object, in r535_curs_init() 133 chan->object.oclass, sizeof(*args), &chan->rm.object); in r535_curs_init() 139 return nvkm_gsp_rm_alloc_wr(&chan->rm.object, args); in r535_curs_init() 161 (chan->disp->rm.client.object.handle & 0x3fff)); in r535_dmac_bind() 184 args = nvkm_gsp_rm_alloc_get(&chan->disp->rm.object, in r535_dmac_init() 186 chan->object.oclass, sizeof(*args), &chan->rm.object); in r535_dmac_init() 193 return nvkm_gsp_rm_alloc_wr(&chan->rm.object, args); in r535_dmac_init() 268 ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, in r535_sor_bl_set() 277 return nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl); in r535_sor_bl_set() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | r535.c | 36 &vmm->rm.client, &vmm->rm.device); in r535_mmu_promote_vmm() 40 args = nvkm_gsp_rm_alloc_get(&vmm->rm.device.object, 0x90f10000, FERMI_VASPACE_A, in r535_mmu_promote_vmm() 41 sizeof(*args), &vmm->rm.object); in r535_mmu_promote_vmm() 47 ret = nvkm_gsp_rm_alloc_wr(&vmm->rm.object, args); in r535_mmu_promote_vmm() 56 &vmm->rm.rsvd); in r535_mmu_promote_vmm() 61 ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.object, in r535_mmu_promote_vmm() 68 ctrl->virtAddrLo = vmm->rm.rsvd->addr; in r535_mmu_promote_vmm() 69 ctrl->virtAddrHi = vmm->rm.rsvd->addr + vmm->rm.rsvd->size - 1; in r535_mmu_promote_vmm() 86 ret = nvkm_gsp_rm_ctrl_wr(&vmm->rm.object, ctrl); in r535_mmu_promote_vmm() 103 struct nvkm_mmu_func *rm; in r535_mmu_new() local [all …]
|
/linux/arch/arm64/net/ |
H A D | bpf_jit.h | 56 #define A64_LS_REG(Rt, Rn, Rm, size, type) \ argument 57 aarch64_insn_gen_load_store_reg(Rt, Rn, Rm, \ 212 #define A64_ADDSUB_SREG(sf, Rd, Rn, Rm, type) \ argument 213 aarch64_insn_gen_add_sub_shifted_reg(Rd, Rn, Rm, 0, \ 215 /* Rd = Rn OP Rm */ 216 #define A64_ADD(sf, Rd, Rn, Rm) A64_ADDSUB_SREG(sf, Rd, Rn, Rm, ADD) argument 217 #define A64_SUB(sf, Rd, Rn, Rm) A64_ADDSUB_SREG(sf, Rd, Rn, Rm, SUB) argument 218 #define A64_SUBS(sf, Rd, Rn, Rm) A64_ADDSUB_SREG(sf, Rd, Rn, Rm, SUB_SETFLAGS) argument 219 /* Rd = -Rm */ 220 #define A64_NEG(sf, Rd, Rm) A64_SUB(sf, Rd, A64_ZR, Rm) argument [all …]
|
/linux/tools/perf/arch/x86/util/ |
H A D | perf_regs.c | 200 * rm[0]: +/-NUM(REG) in arch_sdt_arg_parse_op() 201 * rm[1]: +/- in arch_sdt_arg_parse_op() 202 * rm[2]: NUM in arch_sdt_arg_parse_op() 203 * rm[3]: ( in arch_sdt_arg_parse_op() 204 * rm[4]: REG in arch_sdt_arg_parse_op() 205 * rm[5]: ) in arch_sdt_arg_parse_op() 207 regmatch_t rm[6]; in arch_sdt_arg_parse_op() local 226 regexec(&sdt_op_regex, old_op, 6, rm, 0) || in arch_sdt_arg_parse_op() 227 rm[4].rm_eo - rm[4].rm_so > SDT_REG_NAME_SIZE) { in arch_sdt_arg_parse_op() 244 if (rm[3].rm_so != rm[3].rm_eo) { in arch_sdt_arg_parse_op() [all …]
|
/linux/sound/soc/codecs/ |
H A D | tscs42xx.h | 114 #define RM(m, b) ((m)<<(b)) macro 133 #define RM_HPVOLL RM(FM_HPVOLL, FB_HPVOLL) 156 #define RM_HPVOLR RM(FM_HPVOLR, FB_HPVOLR) 179 #define RM_SPKVOLL RM(FM_SPKVOLL, FB_SPKVOLL) 204 #define RM_SPKVOLR RM(FM_SPKVOLR, FB_SPKVOLR) 229 #define RM_DACVOLL RM(FM_DACVOLL, FB_DACVOLL) 254 #define RM_DACVOLR RM(FM_DACVOLR, FB_DACVOLR) 279 #define RM_ADCVOLL RM(FM_ADCVOLL, FB_ADCVOLL) 304 #define RM_ADCVOLR RM(FM_ADCVOLR, FB_ADCVOLR) 337 RM(FM_INVOLL_INMUTEL, FB_INVOLL_INMUTEL) [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
H A D | r535.c | 69 nvkm_gsp_rm_free(&chan->rm.object); in r535_chan_ramfc_clear() 71 dma_free_coherent(fifo->engine.subdev.device->dev, fifo->rm.mthdbuf_size, in r535_chan_ramfc_clear() 72 chan->rm.mthdbuf.ptr, chan->rm.mthdbuf.addr); in r535_chan_ramfc_clear() 74 nvkm_cgrp_vctx_put(chan->cgrp, &chan->rm.grctx); in r535_chan_ramfc_clear() 105 chan->rm.mthdbuf.ptr = dma_alloc_coherent(fifo->engine.subdev.device->dev, in r535_chan_ramfc_write() 106 fifo->rm.mthdbuf_size, in r535_chan_ramfc_write() 107 &chan->rm.mthdbuf.addr, GFP_KERNEL); in r535_chan_ramfc_write() 108 if (!chan->rm.mthdbuf.ptr) in r535_chan_ramfc_write() 111 args = nvkm_gsp_rm_alloc_get(&chan->vmm->rm.device.object, 0xf1f00000 | chan->id, in r535_chan_ramfc_write() 113 &chan->rm.object); in r535_chan_ramfc_write() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/engine/ofa/ |
H A D | r535.c | 34 struct nvkm_gsp_object rm; member 42 nvkm_gsp_rm_free(&obj->rm); in r535_ofa_obj_dtor() 65 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_ofa_obj_ctor() 66 sizeof(*args), &obj->rm); in r535_ofa_obj_ctor() 72 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_ofa_obj_ctor() 86 struct nvkm_engine_func *rm; in r535_ofa_new() local 91 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_ofa_new() 94 rm->dtor = r535_ofa_dtor; in r535_ofa_new() 96 rm->sclass[i].minver = hw->sclass[i].minver; in r535_ofa_new() 97 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_ofa_new() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/engine/nvjpg/ |
H A D | r535.c | 33 struct nvkm_gsp_object rm; member 41 nvkm_gsp_rm_free(&obj->rm); in r535_nvjpg_obj_dtor() 64 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_nvjpg_obj_ctor() 65 sizeof(*args), &obj->rm); in r535_nvjpg_obj_ctor() 72 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_nvjpg_obj_ctor() 86 struct nvkm_engine_func *rm; in r535_nvjpg_new() local 91 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_nvjpg_new() 94 rm->dtor = r535_nvjpg_dtor; in r535_nvjpg_new() 96 rm->sclass[i].minver = hw->sclass[i].minver; in r535_nvjpg_new() 97 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_nvjpg_new() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/engine/nvdec/ |
H A D | r535.c | 33 struct nvkm_gsp_object rm; member 41 nvkm_gsp_rm_free(&obj->rm); in r535_nvdec_obj_dtor() 64 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_nvdec_obj_ctor() 65 sizeof(*args), &obj->rm); in r535_nvdec_obj_ctor() 72 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_nvdec_obj_ctor() 88 struct nvkm_engine_func *rm; in r535_nvdec_new() local 93 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_nvdec_new() 96 rm->dtor = r535_nvdec_dtor; in r535_nvdec_new() 98 rm->sclass[i].minver = hw->sclass[i].minver; in r535_nvdec_new() 99 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_nvdec_new() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/engine/ce/ |
H A D | r535.c | 34 struct nvkm_gsp_object rm; member 42 nvkm_gsp_rm_free(&obj->rm); in r535_ce_obj_dtor() 65 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_ce_obj_ctor() 66 sizeof(*args), &obj->rm); in r535_ce_obj_ctor() 73 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_ce_obj_ctor() 87 struct nvkm_engine_func *rm; in r535_ce_new() local 92 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_ce_new() 95 rm->dtor = r535_ce_dtor; in r535_ce_new() 97 rm->sclass[i].minver = hw->sclass[i].minver; in r535_ce_new() 98 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_ce_new() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/engine/nvenc/ |
H A D | r535.c | 33 struct nvkm_gsp_object rm; member 41 nvkm_gsp_rm_free(&obj->rm); in r535_nvenc_obj_dtor() 64 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_nvenc_obj_ctor() 65 sizeof(*args), &obj->rm); in r535_nvenc_obj_ctor() 72 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_nvenc_obj_ctor() 88 struct nvkm_engine_func *rm; in r535_nvenc_new() local 93 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_nvenc_new() 96 rm->dtor = r535_nvenc_dtor; in r535_nvenc_new() 98 rm->sclass[i].minver = hw->sclass[i].minver; in r535_nvenc_new() 99 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_nvenc_new() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/bar/ |
H A D | r535.c | 81 vmm->rm.bar2_pdb = gsp->bar.rm_bar2_pdb; in r535_bar_bar2_init() 151 struct nvkm_bar_func *rm; in r535_bar_new_() local 155 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_bar_new_() 158 rm->dtor = r535_bar_dtor; in r535_bar_new_() 159 rm->oneinit = hw->oneinit; in r535_bar_new_() 160 rm->bar1.init = r535_bar_bar1_init; in r535_bar_new_() 161 rm->bar1.fini = r535_bar_bar1_fini; in r535_bar_new_() 162 rm->bar1.wait = r535_bar_bar1_wait; in r535_bar_new_() 163 rm->bar1.vmm = hw->bar1.vmm; in r535_bar_new_() 164 rm->bar2.init = r535_bar_bar2_init; in r535_bar_new_() [all …]
|
/linux/tools/perf/arch/powerpc/util/ |
H A D | perf_regs.c | 138 regmatch_t rm[5]; in arch_sdt_arg_parse_op() local 151 if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) { in arch_sdt_arg_parse_op() 155 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in arch_sdt_arg_parse_op() 162 (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so); in arch_sdt_arg_parse_op() 163 } else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) { in arch_sdt_arg_parse_op() 168 prefix = (rm[1].rm_so == -1) ? '+' : '-'; in arch_sdt_arg_parse_op() 171 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in arch_sdt_arg_parse_op() 172 new_len += (int)(rm[4].rm_eo - rm[4].rm_so); in arch_sdt_arg_parse_op() 179 (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so, in arch_sdt_arg_parse_op() 180 (int)(rm[4].rm_eo - rm[4].rm_so), old_op + rm[4].rm_so); in arch_sdt_arg_parse_op()
|
/linux/tools/perf/arch/arm64/util/ |
H A D | perf_regs.c | 98 regmatch_t rm[5]; in arch_sdt_arg_parse_op() local 104 if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) { in arch_sdt_arg_parse_op() 107 new_len += (int)(rm[1].rm_eo - rm[1].rm_so); in arch_sdt_arg_parse_op() 114 (int)(rm[1].rm_eo - rm[1].rm_so), old_op + rm[1].rm_so); in arch_sdt_arg_parse_op() 115 } else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) { in arch_sdt_arg_parse_op() 120 if (rm[2].rm_so == -1) in arch_sdt_arg_parse_op() 123 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in arch_sdt_arg_parse_op() 129 if (rm[2].rm_so == -1) in arch_sdt_arg_parse_op() 133 (int)(rm[2].rm_eo - rm[2].rm_so), in arch_sdt_arg_parse_op() 134 old_op + rm[2].rm_so); in arch_sdt_arg_parse_op()
|
/linux/arch/mips/math-emu/ |
H A D | dp_mul.c | 16 u64 rm; in ieee754dp_mul() local 104 /* rm = xm * ym, re = xe+ye basically */ in ieee754dp_mul() 116 * Multiply 64 bits xm, ym to give high 64 bits rm with stickness. in ieee754dp_mul() 143 rm = hrm | (lrm != 0); in ieee754dp_mul() 148 if ((s64) rm < 0) { in ieee754dp_mul() 149 rm = (rm >> (64 - (DP_FBITS + 1 + 3))) | in ieee754dp_mul() 150 ((rm << (DP_FBITS + 1 + 3)) != 0); in ieee754dp_mul() 153 rm = (rm >> (64 - (DP_FBITS + 1 + 3 + 1))) | in ieee754dp_mul() 154 ((rm << (DP_FBITS + 1 + 3 + 1)) != 0); in ieee754dp_mul() 156 assert(rm & (DP_HIDDEN_BIT << 3)); in ieee754dp_mul() [all …]
|