Home
last modified time | relevance | path

Searched refs:rm (Results 1 – 25 of 472) sorted by relevance

12345678910>>...19

/linux/drivers/macintosh/
H A Drack-meter.c51 struct rackmeter *rm; member
95 static void rackmeter_setup_i2s(struct rackmeter *rm) in rackmeter_setup_i2s() argument
97 struct macio_chip *macio = rm->mdev->bus->chip; in rackmeter_setup_i2s()
106 pmac_call_feature(PMAC_FTR_SOUND_CHIP_ENABLE, rm->i2s, 0, 1); in rackmeter_setup_i2s()
121 out_le32(rm->i2s_regs + 0x10, 0x01fa0000); in rackmeter_setup_i2s()
122 (void)in_le32(rm->i2s_regs + 0x10); in rackmeter_setup_i2s()
132 static void rackmeter_set_default_pattern(struct rackmeter *rm) in rackmeter_set_default_pattern() argument
138 rm->ubuf[i] = (i & 1) * 255; in rackmeter_set_default_pattern()
140 rm->ubuf[i] = ((~i) & 1) * 255; in rackmeter_set_default_pattern()
144 static void rackmeter_do_pause(struct rackmeter *rm, int pause) in rackmeter_do_pause() argument
[all …]
/linux/arch/arm/net/
H A Dbpf_jit_32.h165 #define _AL3_R(op, rd, rn, rm) ((op ## _R) | (rd) << 12 | (rn) << 16 | (rm)) argument
171 #define ARM_ADD_R(rd, rn, rm) _AL3_R(ARM_INST_ADD, rd, rn, rm) argument
172 #define ARM_ADDS_R(rd, rn, rm) _AL3_R(ARM_INST_ADDS, rd, rn, rm) argument
175 #define ARM_ADC_R(rd, rn, rm) _AL3_R(ARM_INST_ADC, rd, rn, rm) argument
178 #define ARM_AND_R(rd, rn, rm) _AL3_R(ARM_INST_AND, rd, rn, rm) argument
179 #define ARM_ANDS_R(rd, rn, rm) _AL3_R(ARM_INST_ANDS, rd, rn, rm) argument
182 #define ARM_BIC_R(rd, rn, rm) _AL3_R(ARM_INST_BIC, rd, rn, rm) argument
186 #define ARM_BX(rm) (ARM_INST_BX | (rm)) argument
187 #define ARM_BLX_R(rm) (ARM_INST_BLX_R | (rm)) argument
189 #define ARM_CMP_R(rn, rm) _AL3_R(ARM_INST_CMP, 0, rn, rm) argument
[all …]
/linux/net/rds/
H A Dmessage.c53 void rds_message_addref(struct rds_message *rm) in rds_message_addref() argument
55 rdsdebug("addref rm %p ref %d\n", rm, refcount_read(&rm->m_refcount)); in rds_message_addref()
56 refcount_inc(&rm->m_refcount); in rds_message_addref()
132 static void rds_message_purge(struct rds_message *rm) in rds_message_purge() argument
138 if (unlikely(test_bit(RDS_MSG_PAGEVEC, &rm->m_flags))) in rds_message_purge()
141 spin_lock_irqsave(&rm->m_rs_lock, flags); in rds_message_purge()
142 znotifier = rm->data.op_mmp_znotifier; in rds_message_purge()
143 rm->data.op_mmp_znotifier = NULL; in rds_message_purge()
146 if (rm->m_rs) { in rds_message_purge()
147 struct rds_sock *rs = rm->m_rs; in rds_message_purge()
[all …]
H A Dsend.c67 struct rds_message *rm, *tmp; in rds_send_path_reset() local
71 rm = cp->cp_xmit_rm; in rds_send_path_reset()
77 rds_message_unmapped(rm); in rds_send_path_reset()
78 rds_message_put(rm); in rds_send_path_reset()
95 list_for_each_entry_safe(rm, tmp, &cp->cp_retrans, m_conn_item) { in rds_send_path_reset()
96 set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags); in rds_send_path_reset()
97 set_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags); in rds_send_path_reset()
190 struct rds_message *rm; in rds_send_xmit() local
252 rm = cp->cp_xmit_rm; in rds_send_xmit()
254 if (!rm) { in rds_send_xmit()
532 rds_send_sndbuf_remove(struct rds_sock * rs,struct rds_message * rm) rds_send_sndbuf_remove() argument
545 rds_send_is_acked(struct rds_message * rm,u64 ack,is_acked_func is_acked) rds_send_is_acked() argument
559 rds_rdma_send_complete(struct rds_message * rm,int status) rds_rdma_send_complete() argument
595 rds_atomic_send_complete(struct rds_message * rm,int status) rds_atomic_send_complete() argument
634 __rds_send_complete(struct rds_sock * rs,struct rds_message * rm,int status) __rds_send_complete() argument
668 struct rds_message *rm; rds_send_remove_from_sock() local
748 struct rds_message *rm, *tmp; rds_send_path_drop_acked() local
783 struct rds_message *rm, *tmp; rds_send_drop_to() local
879 rds_send_queue_rm(struct rds_sock * rs,struct rds_connection * conn,struct rds_conn_path * cp,struct rds_message * rm,__be16 sport,__be16 dport,int * queued) rds_send_queue_rm() argument
1030 rds_cmsg_zcopy(struct rds_sock * rs,struct rds_message * rm,struct cmsghdr * cmsg) rds_cmsg_zcopy() argument
1043 rds_cmsg_send(struct rds_sock * rs,struct rds_message * rm,struct msghdr * msg,int * allocated_mr,struct rds_iov_vector_arr * vct) rds_cmsg_send() argument
1134 struct rds_message *rm = NULL; rds_sendmsg() local
1475 struct rds_message *rm; rds_send_probe() local
[all...]
H A Dtcp_send.c70 int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm, in rds_tcp_xmit() argument
73 struct rds_conn_path *cp = rm->m_inc.i_conn_path; in rds_tcp_xmit()
86 rm->m_ack_seq = tc->t_last_sent_nxt + in rds_tcp_xmit()
88 be32_to_cpu(rm->m_inc.i_hdr.h_len) - 1; in rds_tcp_xmit()
90 set_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags); in rds_tcp_xmit()
91 tc->t_last_expected_una = rm->m_ack_seq + 1; in rds_tcp_xmit()
93 if (test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags)) in rds_tcp_xmit()
94 rm->m_inc.i_hdr.h_flags |= RDS_FLAG_RETRANSMITTED; in rds_tcp_xmit()
96 rdsdebug("rm %p tcp nxt %u ack_seq %llu\n", in rds_tcp_xmit()
97 rm, rds_tcp_write_se in rds_tcp_xmit()
172 rds_tcp_is_acked(struct rds_message * rm,uint64_t ack) rds_tcp_is_acked() argument
[all...]
H A Drdma.c605 int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm,
610 struct rm_rdma_op *op = &rm->rdma; in rds_cmsg_rdma_args()
620 || rm->rdma.op_active) in rds_cmsg_rdma_args()
666 op->op_sg = rds_message_alloc_sgs(rm, nr_pages); in rds_cmsg_rdma_args()
804 int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm,
813 rm->m_rdma_cookie != 0) in rds_cmsg_rdma_dest()
816 memcpy(&rm->m_rdma_cookie, CMSG_DATA(cmsg), sizeof(rm->m_rdma_cookie)); in rds_cmsg_rdma_dest()
823 r_key = rds_rdma_cookie_key(rm->m_rdma_cookie); in rds_cmsg_rdma_dest()
836 rm in rds_cmsg_rdma_dest()
609 rds_cmsg_rdma_args(struct rds_sock * rs,struct rds_message * rm,struct cmsghdr * cmsg,struct rds_iov_vector * vec) rds_cmsg_rdma_args() argument
808 rds_cmsg_rdma_dest(struct rds_sock * rs,struct rds_message * rm,struct cmsghdr * cmsg) rds_cmsg_rdma_dest() argument
851 rds_cmsg_rdma_map(struct rds_sock * rs,struct rds_message * rm,struct cmsghdr * cmsg) rds_cmsg_rdma_map() argument
865 rds_cmsg_atomic(struct rds_sock * rs,struct rds_message * rm,struct cmsghdr * cmsg) rds_cmsg_atomic() argument
[all...]
H A Dib_send.c48 static void rds_ib_send_complete(struct rds_message *rm, in rds_ib_send_complete() argument
50 void (*complete)(struct rds_message *rm, int status)) in rds_ib_send_complete() argument
70 complete(rm, notify_status); in rds_ib_send_complete()
146 * Returns the rm for no good reason other than it is unobtainable
154 struct rds_message *rm = NULL; in rds_ib_send_unmap_op() local
160 rm = container_of(send->s_op, struct rds_message, data); in rds_ib_send_unmap_op()
167 rm = container_of(send->s_op, struct rds_message, rdma); in rds_ib_send_unmap_op()
174 rm = container_of(send->s_op, struct rds_message, atomic); in rds_ib_send_unmap_op()
187 return rm; in rds_ib_send_unmap_op()
245 struct rds_message *rm in rds_ib_send_cqe_handler() local
484 rds_ib_xmit(struct rds_connection * conn,struct rds_message * rm,unsigned int hdr_off,unsigned int sg,unsigned int off) rds_ib_xmit() argument
[all...]
H A Dloop.c75 static int rds_loop_xmit(struct rds_connection *conn, struct rds_message *rm, in rds_loop_xmit() argument
79 struct scatterlist *sgp = &rm->data.op_sg[sg]; in rds_loop_xmit()
81 be32_to_cpu(rm->m_inc.i_hdr.h_len); in rds_loop_xmit()
84 if (rm->m_inc.i_hdr.h_flags & RDS_FLAG_CONG_BITMAP) { in rds_loop_xmit()
92 rds_inc_init(&rm->m_inc, conn, &conn->c_laddr); in rds_loop_xmit()
94 rds_message_addref(rm); in rds_loop_xmit()
96 rds_recv_incoming(conn, &conn->c_laddr, &conn->c_faddr, &rm->m_inc, in rds_loop_xmit()
99 rds_send_drop_acked(conn, be64_to_cpu(rm->m_inc.i_hdr.h_sequence), in rds_loop_xmit()
102 rds_inc_put(&rm->m_inc); in rds_loop_xmit()
113 struct rds_message *rm = container_of(inc, struct rds_message, m_inc); in rds_loop_inc_free() local
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/
H A Dvmm.c30 if (vmm->rm.external) { in r535_mmu_vaspace_del()
33 ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.device.object, in r535_mmu_vaspace_del()
37 ctrl->hVASpace = vmm->rm.object.handle; in r535_mmu_vaspace_del()
39 WARN_ON(nvkm_gsp_rm_ctrl_wr(&vmm->rm.device.object, ctrl)); in r535_mmu_vaspace_del()
42 vmm->rm.external = false; in r535_mmu_vaspace_del()
45 nvkm_gsp_rm_free(&vmm->rm.object); in r535_mmu_vaspace_del()
46 nvkm_gsp_device_dtor(&vmm->rm.device); in r535_mmu_vaspace_del()
47 nvkm_gsp_client_dtor(&vmm->rm.client); in r535_mmu_vaspace_del()
49 nvkm_vmm_put(vmm, &vmm->rm.rsvd); in r535_mmu_vaspace_del()
59 &vmm->rm.client, &vmm->rm.device); in r535_mmu_vaspace_new()
[all …]
H A Dfifo.c44 struct nvkm_gsp *gsp = chan->rm.object.client->gsp; in r535_chan_doorbell_handle()
46 return gsp->rm->gpu->fifo.chan.doorbell_handle(chan); in r535_chan_doorbell_handle()
64 nvkm_gsp_rm_free(&chan->rm.object); in r535_chan_ramfc_clear()
66 dma_free_coherent(fifo->engine.subdev.device->dev, fifo->rm.mthdbuf_size, in r535_chan_ramfc_clear()
67 chan->rm.mthdbuf.ptr, chan->rm.mthdbuf.addr); in r535_chan_ramfc_clear()
69 nvkm_cgrp_vctx_put(chan->cgrp, &chan->rm.grctx); in r535_chan_ramfc_clear()
120 args->hVASpace = vmm->rm.object.handle; in r535_chan_alloc()
139 args->mthdbufMem.size = fifo->rm.mthdbuf_size; in r535_chan_alloc()
159 const struct nvkm_rm_api *rmapi = device->gsp->rm->api; in r535_chan_ramfc_write()
177 chan->rm.mthdbuf.ptr = dma_alloc_coherent(fifo->engine.subdev.device->dev, in r535_chan_ramfc_write()
[all …]
H A Dbar.c98 vmm->rm.bar2_pdb = gsp->bar.rm_bar2_pdb; in r535_bar_bar2_init()
168 struct nvkm_bar_func *rm; in r535_bar_new_() local
172 if (!(rm = kzalloc_obj(*rm))) in r535_bar_new_()
175 rm->dtor = r535_bar_dtor; in r535_bar_new_()
176 rm->oneinit = hw->oneinit; in r535_bar_new_()
177 rm->bar1.init = r535_bar_bar1_init; in r535_bar_new_()
178 rm->bar1.fini = r535_bar_bar1_fini; in r535_bar_new_()
179 rm->bar1.wait = r535_bar_bar1_wait; in r535_bar_new_()
180 rm->bar1.vmm = hw->bar1.vmm; in r535_bar_new_()
181 rm->bar2.init = r535_bar_bar2_init; in r535_bar_new_()
[all …]
/linux/drivers/gpu/drm/msm/disp/dpu1/
H A Ddpu_rm.c41 struct dpu_rm *rm, in dpu_rm_init() argument
48 if (!rm || !cat || !mmio) { in dpu_rm_init()
54 memset(rm, 0, sizeof(*rm)); in dpu_rm_init()
56 rm->has_legacy_ctls = (cat->mdss_ver->core_major_ver < 5); in dpu_rm_init()
69 rm->mixer_blks[lm->id - LM_0] = &hw->base; in dpu_rm_init()
83 rm->merge_3d_blks[merge_3d->id - MERGE_3D_0] = &hw->base; in dpu_rm_init()
98 hw->merge_3d = to_dpu_hw_merge_3d(rm->merge_3d_blks[pp->merge_3d - MERGE_3D_0]); in dpu_rm_init()
99 rm->pingpong_blks[pp->id - PINGPONG_0] = &hw->base; in dpu_rm_init()
112 rm->hw_intf[intf->id - INTF_0] = hw; in dpu_rm_init()
125 rm->hw_wb[wb->id - WB_0] = hw; in dpu_rm_init()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/
H A Dengine.c19 struct nvkm_gsp_object rm; member
27 nvkm_gsp_rm_free(&obj->rm); in nvkm_rm_engine_obj_dtor()
40 struct nvkm_rm *rm = chan->client->gsp->rm; in nvkm_rm_engine_obj_new() local
53 ret = rm->api->ce->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new()
56 ret = nvkm_gsp_rm_alloc(chan, handle, class, 0, &obj->rm); in nvkm_rm_engine_obj_new()
59 ret = rm->api->nvdec->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new()
62 ret = rm->api->nvenc->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new()
65 ret = rm->api->nvjpg->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new()
68 ret = rm->api->ofa->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new()
92 return nvkm_rm_engine_obj_new(&chan->rm.object, chan->id, oclass, pobject); in nvkm_rm_engine_obj_ctor()
[all …]
H A Dgr.c16 return nvkm_rm_engine_obj_new(&chan->chan->rm.object, chan->chan->id, oclass, pobject); in nvkm_rm_gr_obj_ctor()
22 struct nvkm_rm *rm = base->engine.subdev.device->gsp->rm; in nvkm_rm_gr_fini() local
25 if (rm->api->gr->scrubber.fini) in nvkm_rm_gr_fini()
26 rm->api->gr->scrubber.fini(gr); in nvkm_rm_gr_fini()
34 struct nvkm_rm *rm = base->engine.subdev.device->gsp->rm; in nvkm_rm_gr_init() local
38 if (rm->api->gr->scrubber.init) { in nvkm_rm_gr_init()
39 ret = rm->api->gr->scrubber.init(gr); in nvkm_rm_gr_init()
48 nvkm_rm_gr_new(struct nvkm_rm *rm) in nvkm_rm_gr_new() argument
51 rm->gpu->gr.class.i2m, in nvkm_rm_gr_new()
52 rm->gpu->gr.class.twod, in nvkm_rm_gr_new()
[all …]
/linux/tools/perf/util/perf-regs-arch/
H A Dperf_regs_x86.c166 regmatch_t rm[6]; in __perf_sdt_arg_parse_op_x86() local
185 regexec(&sdt_op_regex, old_op, 6, rm, 0) || in __perf_sdt_arg_parse_op_x86()
186 rm[4].rm_eo - rm[4].rm_so > SDT_REG_NAME_SIZE) { in __perf_sdt_arg_parse_op_x86()
203 if (rm[3].rm_so != rm[3].rm_eo) { in __perf_sdt_arg_parse_op_x86()
204 if (rm[1].rm_so != rm[1].rm_eo) in __perf_sdt_arg_parse_op_x86()
205 prefix[0] = *(old_op + rm[1].rm_so); in __perf_sdt_arg_parse_op_x86()
206 else if (rm[2].rm_so != rm[2].rm_eo) in __perf_sdt_arg_parse_op_x86()
213 sdt_rename_register(old_op + rm[4].rm_so, rm[4].rm_eo - rm[4].rm_so, in __perf_sdt_arg_parse_op_x86()
218 (rm[2].rm_eo - rm[2].rm_so) + in __perf_sdt_arg_parse_op_x86()
219 (rm[3].rm_eo - rm[3].rm_so) + in __perf_sdt_arg_parse_op_x86()
[all …]
H A Dperf_regs_powerpc.c77 regmatch_t rm[5]; in __perf_sdt_arg_parse_op_powerpc() local
90 if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) { in __perf_sdt_arg_parse_op_powerpc()
94 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in __perf_sdt_arg_parse_op_powerpc()
101 (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so); in __perf_sdt_arg_parse_op_powerpc()
102 } else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) { in __perf_sdt_arg_parse_op_powerpc()
107 prefix = (rm[1].rm_so == -1) ? '+' : '-'; in __perf_sdt_arg_parse_op_powerpc()
110 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in __perf_sdt_arg_parse_op_powerpc()
111 new_len += (int)(rm[4].rm_eo - rm[4].rm_so); in __perf_sdt_arg_parse_op_powerpc()
118 (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so, in __perf_sdt_arg_parse_op_powerpc()
119 (int)(rm[4].rm_eo - rm[4].rm_so), old_op + rm[4].rm_so); in __perf_sdt_arg_parse_op_powerpc()
H A Dperf_regs_aarch64.c62 regmatch_t rm[5]; in __perf_sdt_arg_parse_op_arm64() local
68 if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) { in __perf_sdt_arg_parse_op_arm64()
71 new_len += (int)(rm[1].rm_eo - rm[1].rm_so); in __perf_sdt_arg_parse_op_arm64()
78 (int)(rm[1].rm_eo - rm[1].rm_so), old_op + rm[1].rm_so); in __perf_sdt_arg_parse_op_arm64()
79 } else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) { in __perf_sdt_arg_parse_op_arm64()
84 if (rm[2].rm_so == -1) in __perf_sdt_arg_parse_op_arm64()
87 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in __perf_sdt_arg_parse_op_arm64()
93 if (rm[2].rm_so == -1) in __perf_sdt_arg_parse_op_arm64()
97 (int)(rm[2].rm_eo - rm[2].rm_so), in __perf_sdt_arg_parse_op_arm64()
98 old_op + rm[2].rm_so); in __perf_sdt_arg_parse_op_arm64()
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/vfn/
H A Dr535.c37 const struct nvkm_rm_gpu *gpu = device->gsp->rm->gpu; in r535_vfn_new()
38 struct nvkm_vfn_func *rm; in r535_vfn_new() local
41 if (!(rm = kzalloc_obj(*rm))) in r535_vfn_new()
44 rm->dtor = r535_vfn_dtor; in r535_vfn_new()
45 rm->intr = &tu102_vfn_intr; in r535_vfn_new()
46 rm->user.addr = 0x030000; in r535_vfn_new()
47 rm->user.size = 0x010000; in r535_vfn_new()
48 rm->user.base.minver = -1; in r535_vfn_new()
49 rm->user.base.maxver = -1; in r535_vfn_new()
50 rm->user.base.oclass = gpu->usermode.class; in r535_vfn_new()
[all …]
/linux/arch/x86/math-emu/
H A Dreg_divide.c28 int FPU_div(int flags, int rm, int control_w) in FPU_div() argument
37 deststnr = rm; in FPU_div()
46 a = (FPU_REG *) rm; in FPU_div()
49 a = &st(rm); in FPU_div()
51 taga = FPU_gettagi(rm); in FPU_div()
58 b = (FPU_REG *) rm; in FPU_div()
61 b = &st(rm); in FPU_div()
63 tagb = FPU_gettagi(rm); in FPU_div()
133 return real_2op_NaN((FPU_REG *) rm, flags & 0x0f, 0, in FPU_div()
141 return real_2op_NaN(st0_ptr, tag, rm, in FPU_div()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/devinit/
H A Dr535.c36 struct nvkm_devinit_func *rm; in r535_devinit_new() local
39 if (!(rm = kzalloc_obj(*rm))) in r535_devinit_new()
42 rm->dtor = r535_devinit_dtor; in r535_devinit_new()
43 rm->post = hw->post; in r535_devinit_new()
44 rm->disable = hw->disable; in r535_devinit_new()
46 ret = nv50_devinit_new_(rm, device, type, inst, pdevinit); in r535_devinit_new()
48 kfree(rm); in r535_devinit_new()
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/
H A Dbase.c94 nvkm_firmware_put(gsp->fws.rm); in nvkm_gsp_dtor_fws()
95 gsp->fws.rm = NULL; in nvkm_gsp_dtor_fws()
107 kfree(gsp->rm); in nvkm_gsp_dtor()
148 if (fwif->rm) { in nvkm_gsp_new_()
151 gsp->rm = kzalloc_obj(*gsp->rm); in nvkm_gsp_new_()
152 if (!gsp->rm) in nvkm_gsp_new_()
155 gsp->rm->device = device; in nvkm_gsp_new_()
156 gsp->rm->gpu = fwif->func->rm.gpu; in nvkm_gsp_new_()
157 gsp->rm->wpr = fwif->rm->wpr; in nvkm_gsp_new_()
158 gsp->rm->api = fwif->rm->api; in nvkm_gsp_new_()
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/fb/
H A Dr535.c71 struct nvkm_fb_func *rm; in r535_fb_new() local
74 if (!(rm = kzalloc_obj(*rm))) in r535_fb_new()
77 rm->dtor = r535_fb_dtor; in r535_fb_new()
78 rm->sysmem.flush_page_init = hw->sysmem.flush_page_init; in r535_fb_new()
79 rm->vidmem.size = hw->vidmem.size; in r535_fb_new()
80 rm->ram_new = r535_fb_ram_new; in r535_fb_new()
82 ret = nvkm_fb_new_(rm, device, type, inst, pfb); in r535_fb_new()
84 kfree(rm); in r535_fb_new()
/linux/arch/mips/math-emu/
H A Dsp_mul.c16 unsigned int rm; in ieee754sp_mul() local
138 rm = hrm | (lrm != 0); in ieee754sp_mul()
143 if ((int) rm < 0) { in ieee754sp_mul()
144 rm = (rm >> (32 - (SP_FBITS + 1 + 3))) | in ieee754sp_mul()
145 ((rm << (SP_FBITS + 1 + 3)) != 0); in ieee754sp_mul()
148 rm = (rm >> (32 - (SP_FBITS + 1 + 3 + 1))) | in ieee754sp_mul()
149 ((rm << (SP_FBITS + 1 + 3 + 1)) != 0); in ieee754sp_mul()
151 assert(rm & (SP_HIDDEN_BIT << 3)); in ieee754sp_mul()
153 return ieee754sp_format(rs, re, rm); in ieee754sp_mul()
H A Ddp_mul.c16 u64 rm; in ieee754dp_mul() local
143 rm = hrm | (lrm != 0); in ieee754dp_mul()
148 if ((s64) rm < 0) { in ieee754dp_mul()
149 rm = (rm >> (64 - (DP_FBITS + 1 + 3))) | in ieee754dp_mul()
150 ((rm << (DP_FBITS + 1 + 3)) != 0); in ieee754dp_mul()
153 rm = (rm >> (64 - (DP_FBITS + 1 + 3 + 1))) | in ieee754dp_mul()
154 ((rm << (DP_FBITS + 1 + 3 + 1)) != 0); in ieee754dp_mul()
156 assert(rm & (DP_HIDDEN_BIT << 3)); in ieee754dp_mul()
158 return ieee754dp_format(rs, re, rm); in ieee754dp_mul()
/linux/arch/arm/probes/kprobes/
H A Dactions-arm.c75 int rm = insn & 0xf; in emulate_ldrdstrd() local
81 register unsigned long rmv asm("r3") = regs->uregs[rm]; in emulate_ldrdstrd()
104 int rm = insn & 0xf; in emulate_ldr() local
109 register unsigned long rmv asm("r3") = regs->uregs[rm]; in emulate_ldr()
135 int rm = insn & 0xf; in emulate_str() local
141 register unsigned long rmv asm("r3") = regs->uregs[rm]; in emulate_str()
161 int rm = insn & 0xf; in emulate_rd12rn16rm0rs8_rwflags() local
167 register unsigned long rmv asm("r3") = (rm == 15) ? pc in emulate_rd12rn16rm0rs8_rwflags()
168 : regs->uregs[rm]; in emulate_rd12rn16rm0rs8_rwflags()
195 int rm = insn & 0xf; in emulate_rd12rn16rm0_rwflags_nopc() local
[all …]

12345678910>>...19