Home
last modified time | relevance | path

Searched full:rm (Results 1 – 25 of 807) sorted by relevance

12345678910>>...33

/linux/drivers/gpu/drm/nouveau/include/nvrm/535.113.01/nvidia/kernel/inc/vgpu/
H A Drpc_global_enums.h4 /* Excerpt of RM headers from https://github.com/NVIDIA/open-gpu-kernel-modules/tree/535.113.01 */
11 X(RM, NOP) // 0
12 X(RM, SET_GUEST_SYSTEM_INFO) // 1
13 X(RM, ALLOC_ROOT) // 2
14 X(RM, ALLOC_DEVICE) // 3 deprecated
15 X(RM, ALLOC_MEMORY) // 4
16 X(RM, ALLOC_CTX_DMA) // 5
17 X(RM, ALLOC_CHANNEL_DMA) // 6
18 X(RM, MAP_MEMORY) // 7
19 X(RM, BIND_CTX_DMA) // 8 deprecated
[all …]
/linux/net/rds/
H A Dmessage.c51 void rds_message_addref(struct rds_message *rm) in rds_message_addref() argument
53 rdsdebug("addref rm %p ref %d\n", rm, refcount_read(&rm->m_refcount)); in rds_message_addref()
54 refcount_inc(&rm->m_refcount); in rds_message_addref()
130 static void rds_message_purge(struct rds_message *rm) in rds_message_purge() argument
135 if (unlikely(test_bit(RDS_MSG_PAGEVEC, &rm->m_flags))) in rds_message_purge()
138 spin_lock_irqsave(&rm->m_rs_lock, flags); in rds_message_purge()
139 if (rm->m_rs) { in rds_message_purge()
140 struct rds_sock *rs = rm->m_rs; in rds_message_purge()
142 if (rm->data.op_mmp_znotifier) { in rds_message_purge()
144 rds_rm_zerocopy_callback(rs, rm->data.op_mmp_znotifier); in rds_message_purge()
[all …]
H A Dsend.c67 struct rds_message *rm, *tmp; in rds_send_path_reset() local
71 rm = cp->cp_xmit_rm; in rds_send_path_reset()
77 rds_message_unmapped(rm); in rds_send_path_reset()
78 rds_message_put(rm); in rds_send_path_reset()
95 list_for_each_entry_safe(rm, tmp, &cp->cp_retrans, m_conn_item) { in rds_send_path_reset()
96 set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags); in rds_send_path_reset()
97 set_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags); in rds_send_path_reset()
139 struct rds_message *rm; in rds_send_xmit() local
201 rm = cp->cp_xmit_rm; in rds_send_xmit()
203 if (!rm) { in rds_send_xmit()
[all …]
H A Dtcp_send.c70 int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm, in rds_tcp_xmit() argument
73 struct rds_conn_path *cp = rm->m_inc.i_conn_path; in rds_tcp_xmit()
86 rm->m_ack_seq = tc->t_last_sent_nxt + in rds_tcp_xmit()
88 be32_to_cpu(rm->m_inc.i_hdr.h_len) - 1; in rds_tcp_xmit()
90 set_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags); in rds_tcp_xmit()
91 tc->t_last_expected_una = rm->m_ack_seq + 1; in rds_tcp_xmit()
93 if (test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags)) in rds_tcp_xmit()
94 rm->m_inc.i_hdr.h_flags |= RDS_FLAG_RETRANSMITTED; in rds_tcp_xmit()
96 rdsdebug("rm %p tcp nxt %u ack_seq %llu\n", in rds_tcp_xmit()
97 rm, rds_tcp_write_seq(tc), in rds_tcp_xmit()
[all …]
H A Drdma.c611 int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm, in rds_cmsg_rdma_args() argument
616 struct rm_rdma_op *op = &rm->rdma; in rds_cmsg_rdma_args()
626 || rm->rdma.op_active) in rds_cmsg_rdma_args()
672 op->op_sg = rds_message_alloc_sgs(rm, nr_pages); in rds_cmsg_rdma_args()
811 int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm, in rds_cmsg_rdma_dest() argument
820 rm->m_rdma_cookie != 0) in rds_cmsg_rdma_dest()
823 memcpy(&rm->m_rdma_cookie, CMSG_DATA(cmsg), sizeof(rm->m_rdma_cookie)); in rds_cmsg_rdma_dest()
830 r_key = rds_rdma_cookie_key(rm->m_rdma_cookie); in rds_cmsg_rdma_dest()
843 rm->rdma.op_rdma_mr = mr; in rds_cmsg_rdma_dest()
851 * in rm->m_rdma_cookie. This causes it to be sent along to the peer
[all …]
H A Dib_send.c48 static void rds_ib_send_complete(struct rds_message *rm, in rds_ib_send_complete() argument
50 void (*complete)(struct rds_message *rm, int status)) in rds_ib_send_complete() argument
70 complete(rm, notify_status); in rds_ib_send_complete()
146 * Returns the rm for no good reason other than it is unobtainable
154 struct rds_message *rm = NULL; in rds_ib_send_unmap_op() local
160 rm = container_of(send->s_op, struct rds_message, data); in rds_ib_send_unmap_op()
167 rm = container_of(send->s_op, struct rds_message, rdma); in rds_ib_send_unmap_op()
174 rm = container_of(send->s_op, struct rds_message, atomic); in rds_ib_send_unmap_op()
187 return rm; in rds_ib_send_unmap_op()
245 struct rds_message *rm = NULL; in rds_ib_send_cqe_handler() local
[all …]
/linux/arch/arm/net/
H A Dbpf_jit_32.h165 #define _AL3_R(op, rd, rn, rm) ((op ## _R) | (rd) << 12 | (rn) << 16 | (rm)) argument
171 #define ARM_ADD_R(rd, rn, rm) _AL3_R(ARM_INST_ADD, rd, rn, rm) argument
172 #define ARM_ADDS_R(rd, rn, rm) _AL3_R(ARM_INST_ADDS, rd, rn, rm) argument
175 #define ARM_ADC_R(rd, rn, rm) _AL3_R(ARM_INST_ADC, rd, rn, rm) argument
178 #define ARM_AND_R(rd, rn, rm) _AL3_R(ARM_INST_AND, rd, rn, rm) argument
179 #define ARM_ANDS_R(rd, rn, rm) _AL3_R(ARM_INST_ANDS, rd, rn, rm) argument
182 #define ARM_BIC_R(rd, rn, rm) _AL3_R(ARM_INST_BIC, rd, rn, rm) argument
186 #define ARM_BX(rm) (ARM_INST_BX | (rm)) argument
187 #define ARM_BLX_R(rm) (ARM_INST_BLX_R | (rm)) argument
189 #define ARM_CMP_R(rn, rm) _AL3_R(ARM_INST_CMP, 0, rn, rm) argument
[all …]
/linux/drivers/macintosh/
H A Drack-meter.c51 struct rackmeter *rm; member
95 static void rackmeter_setup_i2s(struct rackmeter *rm) in rackmeter_setup_i2s() argument
97 struct macio_chip *macio = rm->mdev->bus->chip; in rackmeter_setup_i2s()
106 pmac_call_feature(PMAC_FTR_SOUND_CHIP_ENABLE, rm->i2s, 0, 1); in rackmeter_setup_i2s()
121 out_le32(rm->i2s_regs + 0x10, 0x01fa0000); in rackmeter_setup_i2s()
122 (void)in_le32(rm->i2s_regs + 0x10); in rackmeter_setup_i2s()
132 static void rackmeter_set_default_pattern(struct rackmeter *rm) in rackmeter_set_default_pattern() argument
138 rm->ubuf[i] = (i & 1) * 255; in rackmeter_set_default_pattern()
140 rm->ubuf[i] = ((~i) & 1) * 255; in rackmeter_set_default_pattern()
144 static void rackmeter_do_pause(struct rackmeter *rm, int pause) in rackmeter_do_pause() argument
[all …]
/linux/drivers/gpu/drm/msm/disp/dpu1/
H A Ddpu_rm.c38 struct dpu_rm *rm, in dpu_rm_init() argument
45 if (!rm || !cat || !mmio) { in dpu_rm_init()
51 memset(rm, 0, sizeof(*rm)); in dpu_rm_init()
64 rm->mixer_blks[lm->id - LM_0] = &hw->base; in dpu_rm_init()
78 rm->merge_3d_blks[merge_3d->id - MERGE_3D_0] = &hw->base; in dpu_rm_init()
93 hw->merge_3d = to_dpu_hw_merge_3d(rm->merge_3d_blks[pp->merge_3d - MERGE_3D_0]); in dpu_rm_init()
94 rm->pingpong_blks[pp->id - PINGPONG_0] = &hw->base; in dpu_rm_init()
107 rm->hw_intf[intf->id - INTF_0] = hw; in dpu_rm_init()
120 rm->hw_wb[wb->id - WB_0] = hw; in dpu_rm_init()
133 rm->ctl_blks[ctl->id - CTL_0] = &hw->base; in dpu_rm_init()
[all …]
H A Ddpu_rm.h44 * @rm: DPU Resource Manager handle
51 struct dpu_rm *rm,
63 * @rm: DPU Resource Manager handle
69 int dpu_rm_reserve(struct dpu_rm *rm,
78 * @rm: DPU Resource Manager handle
88 int dpu_rm_get_assigned_resources(struct dpu_rm *rm,
93 * dpu_rm_print_state - output the RM private state
102 * @rm: DPU Resource Manager handle
105 static inline struct dpu_hw_intf *dpu_rm_get_intf(struct dpu_rm *rm, enum dpu_intf intf_idx) in dpu_rm_get_intf() argument
107 return rm->hw_intf[intf_idx - INTF_0]; in dpu_rm_get_intf()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/disp/
H A Dr535.c77 nvkm_gsp_rm_free(&chan->rm.object); in r535_chan_fini()
131 args = nvkm_gsp_rm_alloc_get(&chan->disp->rm.object, in r535_curs_init()
133 chan->object.oclass, sizeof(*args), &chan->rm.object); in r535_curs_init()
139 return nvkm_gsp_rm_alloc_wr(&chan->rm.object, args); in r535_curs_init()
161 (chan->disp->rm.client.object.handle & 0x3fff)); in r535_dmac_bind()
184 args = nvkm_gsp_rm_alloc_get(&chan->disp->rm.object, in r535_dmac_init()
186 chan->object.oclass, sizeof(*args), &chan->rm.object); in r535_dmac_init()
193 return nvkm_gsp_rm_alloc_wr(&chan->rm.object, args); in r535_dmac_init()
268 ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, in r535_sor_bl_set()
277 return nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl); in r535_sor_bl_set()
[all …]
/linux/arch/arm64/net/
H A Dbpf_jit.h56 #define A64_LS_REG(Rt, Rn, Rm, size, type) \ argument
57 aarch64_insn_gen_load_store_reg(Rt, Rn, Rm, \
212 #define A64_ADDSUB_SREG(sf, Rd, Rn, Rm, type) \ argument
213 aarch64_insn_gen_add_sub_shifted_reg(Rd, Rn, Rm, 0, \
215 /* Rd = Rn OP Rm */
216 #define A64_ADD(sf, Rd, Rn, Rm) A64_ADDSUB_SREG(sf, Rd, Rn, Rm, ADD) argument
217 #define A64_SUB(sf, Rd, Rn, Rm) A64_ADDSUB_SREG(sf, Rd, Rn, Rm, SUB) argument
218 #define A64_SUBS(sf, Rd, Rn, Rm) A64_ADDSUB_SREG(sf, Rd, Rn, Rm, SUB_SETFLAGS) argument
219 /* Rd = -Rm */
220 #define A64_NEG(sf, Rd, Rm) A64_SUB(sf, Rd, A64_ZR, Rm) argument
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
H A Dr535.c36 &vmm->rm.client, &vmm->rm.device); in r535_mmu_promote_vmm()
40 args = nvkm_gsp_rm_alloc_get(&vmm->rm.device.object, 0x90f10000, FERMI_VASPACE_A, in r535_mmu_promote_vmm()
41 sizeof(*args), &vmm->rm.object); in r535_mmu_promote_vmm()
47 ret = nvkm_gsp_rm_alloc_wr(&vmm->rm.object, args); in r535_mmu_promote_vmm()
56 &vmm->rm.rsvd); in r535_mmu_promote_vmm()
61 ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.object, in r535_mmu_promote_vmm()
68 ctrl->virtAddrLo = vmm->rm.rsvd->addr; in r535_mmu_promote_vmm()
69 ctrl->virtAddrHi = vmm->rm.rsvd->addr + vmm->rm.rsvd->size - 1; in r535_mmu_promote_vmm()
86 ret = nvkm_gsp_rm_ctrl_wr(&vmm->rm.object, ctrl); in r535_mmu_promote_vmm()
103 struct nvkm_mmu_func *rm; in r535_mmu_new() local
[all …]
/linux/tools/perf/arch/x86/util/
H A Dperf_regs.c200 * rm[0]: +/-NUM(REG) in arch_sdt_arg_parse_op()
201 * rm[1]: +/- in arch_sdt_arg_parse_op()
202 * rm[2]: NUM in arch_sdt_arg_parse_op()
203 * rm[3]: ( in arch_sdt_arg_parse_op()
204 * rm[4]: REG in arch_sdt_arg_parse_op()
205 * rm[5]: ) in arch_sdt_arg_parse_op()
207 regmatch_t rm[6]; in arch_sdt_arg_parse_op() local
226 regexec(&sdt_op_regex, old_op, 6, rm, 0) || in arch_sdt_arg_parse_op()
227 rm[4].rm_eo - rm[4].rm_so > SDT_REG_NAME_SIZE) { in arch_sdt_arg_parse_op()
244 if (rm[3].rm_so != rm[3].rm_eo) { in arch_sdt_arg_parse_op()
[all …]
/linux/sound/soc/codecs/
H A Dtscs42xx.h114 #define RM(m, b) ((m)<<(b)) macro
133 #define RM_HPVOLL RM(FM_HPVOLL, FB_HPVOLL)
156 #define RM_HPVOLR RM(FM_HPVOLR, FB_HPVOLR)
179 #define RM_SPKVOLL RM(FM_SPKVOLL, FB_SPKVOLL)
204 #define RM_SPKVOLR RM(FM_SPKVOLR, FB_SPKVOLR)
229 #define RM_DACVOLL RM(FM_DACVOLL, FB_DACVOLL)
254 #define RM_DACVOLR RM(FM_DACVOLR, FB_DACVOLR)
279 #define RM_ADCVOLL RM(FM_ADCVOLL, FB_ADCVOLL)
304 #define RM_ADCVOLR RM(FM_ADCVOLR, FB_ADCVOLR)
337 RM(FM_INVOLL_INMUTEL, FB_INVOLL_INMUTEL)
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
H A Dr535.c69 nvkm_gsp_rm_free(&chan->rm.object); in r535_chan_ramfc_clear()
71 dma_free_coherent(fifo->engine.subdev.device->dev, fifo->rm.mthdbuf_size, in r535_chan_ramfc_clear()
72 chan->rm.mthdbuf.ptr, chan->rm.mthdbuf.addr); in r535_chan_ramfc_clear()
74 nvkm_cgrp_vctx_put(chan->cgrp, &chan->rm.grctx); in r535_chan_ramfc_clear()
105 chan->rm.mthdbuf.ptr = dma_alloc_coherent(fifo->engine.subdev.device->dev, in r535_chan_ramfc_write()
106 fifo->rm.mthdbuf_size, in r535_chan_ramfc_write()
107 &chan->rm.mthdbuf.addr, GFP_KERNEL); in r535_chan_ramfc_write()
108 if (!chan->rm.mthdbuf.ptr) in r535_chan_ramfc_write()
111 args = nvkm_gsp_rm_alloc_get(&chan->vmm->rm.device.object, 0xf1f00000 | chan->id, in r535_chan_ramfc_write()
113 &chan->rm.object); in r535_chan_ramfc_write()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/ofa/
H A Dr535.c34 struct nvkm_gsp_object rm; member
42 nvkm_gsp_rm_free(&obj->rm); in r535_ofa_obj_dtor()
65 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_ofa_obj_ctor()
66 sizeof(*args), &obj->rm); in r535_ofa_obj_ctor()
72 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_ofa_obj_ctor()
86 struct nvkm_engine_func *rm; in r535_ofa_new() local
91 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_ofa_new()
94 rm->dtor = r535_ofa_dtor; in r535_ofa_new()
96 rm->sclass[i].minver = hw->sclass[i].minver; in r535_ofa_new()
97 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_ofa_new()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/nvjpg/
H A Dr535.c33 struct nvkm_gsp_object rm; member
41 nvkm_gsp_rm_free(&obj->rm); in r535_nvjpg_obj_dtor()
64 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_nvjpg_obj_ctor()
65 sizeof(*args), &obj->rm); in r535_nvjpg_obj_ctor()
72 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_nvjpg_obj_ctor()
86 struct nvkm_engine_func *rm; in r535_nvjpg_new() local
91 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_nvjpg_new()
94 rm->dtor = r535_nvjpg_dtor; in r535_nvjpg_new()
96 rm->sclass[i].minver = hw->sclass[i].minver; in r535_nvjpg_new()
97 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_nvjpg_new()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/nvdec/
H A Dr535.c33 struct nvkm_gsp_object rm; member
41 nvkm_gsp_rm_free(&obj->rm); in r535_nvdec_obj_dtor()
64 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_nvdec_obj_ctor()
65 sizeof(*args), &obj->rm); in r535_nvdec_obj_ctor()
72 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_nvdec_obj_ctor()
88 struct nvkm_engine_func *rm; in r535_nvdec_new() local
93 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_nvdec_new()
96 rm->dtor = r535_nvdec_dtor; in r535_nvdec_new()
98 rm->sclass[i].minver = hw->sclass[i].minver; in r535_nvdec_new()
99 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_nvdec_new()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/ce/
H A Dr535.c34 struct nvkm_gsp_object rm; member
42 nvkm_gsp_rm_free(&obj->rm); in r535_ce_obj_dtor()
65 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_ce_obj_ctor()
66 sizeof(*args), &obj->rm); in r535_ce_obj_ctor()
73 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_ce_obj_ctor()
87 struct nvkm_engine_func *rm; in r535_ce_new() local
92 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_ce_new()
95 rm->dtor = r535_ce_dtor; in r535_ce_new()
97 rm->sclass[i].minver = hw->sclass[i].minver; in r535_ce_new()
98 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_ce_new()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/nvenc/
H A Dr535.c33 struct nvkm_gsp_object rm; member
41 nvkm_gsp_rm_free(&obj->rm); in r535_nvenc_obj_dtor()
64 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_nvenc_obj_ctor()
65 sizeof(*args), &obj->rm); in r535_nvenc_obj_ctor()
72 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_nvenc_obj_ctor()
88 struct nvkm_engine_func *rm; in r535_nvenc_new() local
93 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_nvenc_new()
96 rm->dtor = r535_nvenc_dtor; in r535_nvenc_new()
98 rm->sclass[i].minver = hw->sclass[i].minver; in r535_nvenc_new()
99 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_nvenc_new()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/bar/
H A Dr535.c81 vmm->rm.bar2_pdb = gsp->bar.rm_bar2_pdb; in r535_bar_bar2_init()
151 struct nvkm_bar_func *rm; in r535_bar_new_() local
155 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_bar_new_()
158 rm->dtor = r535_bar_dtor; in r535_bar_new_()
159 rm->oneinit = hw->oneinit; in r535_bar_new_()
160 rm->bar1.init = r535_bar_bar1_init; in r535_bar_new_()
161 rm->bar1.fini = r535_bar_bar1_fini; in r535_bar_new_()
162 rm->bar1.wait = r535_bar_bar1_wait; in r535_bar_new_()
163 rm->bar1.vmm = hw->bar1.vmm; in r535_bar_new_()
164 rm->bar2.init = r535_bar_bar2_init; in r535_bar_new_()
[all …]
/linux/tools/perf/arch/powerpc/util/
H A Dperf_regs.c138 regmatch_t rm[5]; in arch_sdt_arg_parse_op() local
151 if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) { in arch_sdt_arg_parse_op()
155 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in arch_sdt_arg_parse_op()
162 (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so); in arch_sdt_arg_parse_op()
163 } else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) { in arch_sdt_arg_parse_op()
168 prefix = (rm[1].rm_so == -1) ? '+' : '-'; in arch_sdt_arg_parse_op()
171 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in arch_sdt_arg_parse_op()
172 new_len += (int)(rm[4].rm_eo - rm[4].rm_so); in arch_sdt_arg_parse_op()
179 (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so, in arch_sdt_arg_parse_op()
180 (int)(rm[4].rm_eo - rm[4].rm_so), old_op + rm[4].rm_so); in arch_sdt_arg_parse_op()
/linux/tools/perf/arch/arm64/util/
H A Dperf_regs.c98 regmatch_t rm[5]; in arch_sdt_arg_parse_op() local
104 if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) { in arch_sdt_arg_parse_op()
107 new_len += (int)(rm[1].rm_eo - rm[1].rm_so); in arch_sdt_arg_parse_op()
114 (int)(rm[1].rm_eo - rm[1].rm_so), old_op + rm[1].rm_so); in arch_sdt_arg_parse_op()
115 } else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) { in arch_sdt_arg_parse_op()
120 if (rm[2].rm_so == -1) in arch_sdt_arg_parse_op()
123 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in arch_sdt_arg_parse_op()
129 if (rm[2].rm_so == -1) in arch_sdt_arg_parse_op()
133 (int)(rm[2].rm_eo - rm[2].rm_so), in arch_sdt_arg_parse_op()
134 old_op + rm[2].rm_so); in arch_sdt_arg_parse_op()
/linux/arch/mips/math-emu/
H A Ddp_mul.c16 u64 rm; in ieee754dp_mul() local
104 /* rm = xm * ym, re = xe+ye basically */ in ieee754dp_mul()
116 * Multiply 64 bits xm, ym to give high 64 bits rm with stickness. in ieee754dp_mul()
143 rm = hrm | (lrm != 0); in ieee754dp_mul()
148 if ((s64) rm < 0) { in ieee754dp_mul()
149 rm = (rm >> (64 - (DP_FBITS + 1 + 3))) | in ieee754dp_mul()
150 ((rm << (DP_FBITS + 1 + 3)) != 0); in ieee754dp_mul()
153 rm = (rm >> (64 - (DP_FBITS + 1 + 3 + 1))) | in ieee754dp_mul()
154 ((rm << (DP_FBITS + 1 + 3 + 1)) != 0); in ieee754dp_mul()
156 assert(rm & (DP_HIDDEN_BIT << 3)); in ieee754dp_mul()
[all …]

12345678910>>...33