Home
last modified time | relevance | path

Searched refs:msgq (Results 1 – 25 of 26) sorted by relevance

12

/linux/drivers/gpu/drm/nouveau/nvkm/falcon/
H A Dmsgq.c26 nvkm_falcon_msgq_open(struct nvkm_falcon_msgq *msgq) in nvkm_falcon_msgq_open() argument
28 spin_lock(&msgq->lock); in nvkm_falcon_msgq_open()
29 msgq->position = nvkm_falcon_rd32(msgq->qmgr->falcon, msgq->tail_reg); in nvkm_falcon_msgq_open()
33 nvkm_falcon_msgq_close(struct nvkm_falcon_msgq *msgq, bool commit) in nvkm_falcon_msgq_close() argument
35 struct nvkm_falcon *falcon = msgq->qmgr->falcon; in nvkm_falcon_msgq_close()
38 nvkm_falcon_wr32(falcon, msgq->tail_reg, msgq->position); in nvkm_falcon_msgq_close()
40 spin_unlock(&msgq->lock); in nvkm_falcon_msgq_close()
44 nvkm_falcon_msgq_empty(struct nvkm_falcon_msgq *msgq) in nvkm_falcon_msgq_empty() argument
46 u32 head = nvkm_falcon_rd32(msgq->qmgr->falcon, msgq->head_reg); in nvkm_falcon_msgq_empty()
47 u32 tail = nvkm_falcon_rd32(msgq->qmgr->falcon, msgq->tail_reg); in nvkm_falcon_msgq_empty()
[all …]
H A DKbuild5 nvkm-y += nvkm/falcon/msgq.o
/linux/drivers/scsi/arm/
H A Dmsgqueue.c22 static struct msgqueue_entry *mqe_alloc(MsgQueue_t *msgq) in mqe_alloc() argument
26 if ((mq = msgq->free) != NULL) in mqe_alloc()
27 msgq->free = mq->next; in mqe_alloc()
38 static void mqe_free(MsgQueue_t *msgq, struct msgqueue_entry *mq) in mqe_free() argument
41 mq->next = msgq->free; in mqe_free()
42 msgq->free = mq; in mqe_free()
51 void msgqueue_initialise(MsgQueue_t *msgq) in msgqueue_initialise() argument
55 msgq->qe = NULL; in msgqueue_initialise()
56 msgq->free = &msgq->entries[0]; in msgqueue_initialise()
59 msgq->entries[i].next = &msgq->entries[i + 1]; in msgqueue_initialise()
[all …]
H A Dmsgqueue.h36 extern void msgqueue_initialise(MsgQueue_t *msgq);
43 extern void msgqueue_free(MsgQueue_t *msgq);
51 extern int msgqueue_msglength(MsgQueue_t *msgq);
60 extern struct message *msgqueue_getmsg(MsgQueue_t *msgq, int msgno);
70 extern int msgqueue_addmsg(MsgQueue_t *msgq, int length, ...);
77 extern void msgqueue_flush(MsgQueue_t *msgq);
/linux/drivers/net/ethernet/brocade/bna/
H A Dbfa_msgq.c93 bfa_wc_down(&cmdq->msgq->init_wc); in cmdq_sm_init_wait_entry()
195 if (!bfa_nw_ioc_mbox_queue(cmdq->msgq->ioc, &cmdq->dbell_mb, in bfa_msgq_cmdq_dbell()
294 if (!bfa_nw_ioc_mbox_queue(cmdq->msgq->ioc, &cmdq->copy_mb, in bfa_msgq_cmdq_copy_rsp()
301 bfa_msgq_cmdq_attach(struct bfa_msgq_cmdq *cmdq, struct bfa_msgq *msgq) in bfa_msgq_cmdq_attach() argument
305 cmdq->msgq = msgq; in bfa_msgq_cmdq_attach()
356 bfa_wc_down(&rspq->msgq->init_wc); in rspq_sm_init_wait_entry()
403 if (!bfa_nw_ioc_is_disabled(rspq->msgq->ioc)) in rspq_sm_dbell_wait_entry()
451 if (!bfa_nw_ioc_mbox_queue(rspq->msgq->ioc, &rspq->dbell_mb, in bfa_msgq_rspq_dbell()
489 bfa_msgq_rspq_attach(struct bfa_msgq_rspq *rspq, struct bfa_msgq *msgq) in bfa_msgq_rspq_attach() argument
492 rspq->msgq = msgq; in bfa_msgq_rspq_attach()
[all …]
H A Dbfa_msgq.h77 struct bfa_msgq *msgq; member
104 struct bfa_msgq *msgq; member
119 void bfa_msgq_memclaim(struct bfa_msgq *msgq, u8 *kva, u64 pa);
120 void bfa_msgq_attach(struct bfa_msgq *msgq, struct bfa_ioc *ioc);
121 void bfa_msgq_regisr(struct bfa_msgq *msgq, enum bfi_mclass mc,
123 void bfa_msgq_cmd_post(struct bfa_msgq *msgq,
125 void bfa_msgq_rsp_copy(struct bfa_msgq *msgq, u8 *buf, size_t buf_len);
H A Dbna_enet.c418 bfa_msgq_cmd_post(&ethport->bna->msgq, &ethport->msgq_cmd); in bna_bfi_ethport_admin_up()
435 bfa_msgq_cmd_post(&ethport->bna->msgq, &ethport->msgq_cmd); in bna_bfi_ethport_admin_down()
456 bfa_msgq_cmd_post(&ethport->bna->msgq, &ethport->msgq_cmd); in bna_bfi_ethport_lpbk_up()
473 bfa_msgq_cmd_post(&ethport->bna->msgq, &ethport->msgq_cmd); in bna_bfi_ethport_lpbk_down()
1171 bfa_msgq_cmd_post(&enet->bna->msgq, &enet->msgq_cmd); in bna_bfi_pause_set()
1630 bfa_msgq_cmd_post(&ioceth->bna->msgq, &ioceth->msgq_cmd); in bna_bfi_attr_get()
1729 bfa_msgq_attach(&bna->msgq, &ioceth->ioc); in bna_ioceth_init()
1730 bfa_msgq_memclaim(&bna->msgq, kva, dma); in bna_ioceth_init()
1731 bfa_msgq_regisr(&bna->msgq, BFI_MC_ENET, bna_msgq_rsp_handler, bna); in bna_ioceth_init()
1858 bfa_msgq_cmd_post(&bna->msgq, &bna->stats_mod.stats_get_cmd); in bna_bfi_stats_get()
H A Dbna_tx_rx.c190 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_ucast_req()
206 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_mcast_add_req()
222 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_mcast_del_req()
237 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_mcast_filter_req()
252 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rx_promisc_req()
277 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rx_vlan_filter_set()
292 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_vlan_strip_enable()
308 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rit_cfg()
328 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rss_cfg()
343 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rss_enable()
[all …]
/linux/drivers/misc/bcm-vk/
H A Dbcm_vk_msg.c87 static u32 msgq_occupied(const struct bcm_vk_msgq __iomem *msgq, in msgq_occupied() argument
92 wr_idx = readl_relaxed(&msgq->wr_idx); in msgq_occupied()
93 rd_idx = readl_relaxed(&msgq->rd_idx); in msgq_occupied()
99 u32 msgq_avail_space(const struct bcm_vk_msgq __iomem *msgq, in msgq_avail_space() argument
102 return (qinfo->q_size - msgq_occupied(msgq, qinfo) - 1); in msgq_avail_space()
410 struct bcm_vk_msgq __iomem *msgq; in bcm_vk_sync_msgq() local
446 msgq = vk->bar[BAR_1] + msgq_off; in bcm_vk_sync_msgq()
468 chan->msgq[j] = msgq; in bcm_vk_sync_msgq()
469 msgq_start = readl_relaxed(&msgq->start); in bcm_vk_sync_msgq()
470 msgq_size = readl_relaxed(&msgq->size); in bcm_vk_sync_msgq()
[all …]
H A Dbcm_vk_msg.h136 struct bcm_vk_msgq __iomem *msgq[VK_MSGQ_MAX_NR]; member
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/
H A Dr535.c93 u32 size, rptr = *gsp->msgq.rptr; in r535_gsp_msgq_wait()
99 if (WARN_ON(!size || size >= gsp->msgq.cnt)) in r535_gsp_msgq_wait()
103 u32 wptr = *gsp->msgq.wptr; in r535_gsp_msgq_wait()
105 used = wptr + gsp->msgq.cnt - rptr; in r535_gsp_msgq_wait()
106 if (used >= gsp->msgq.cnt) in r535_gsp_msgq_wait()
107 used -= gsp->msgq.cnt; in r535_gsp_msgq_wait()
117 mqe = (void *)((u8 *)gsp->shm.msgq.ptr + 0x1000 + rptr * 0x1000); in r535_gsp_msgq_wait()
128 len = ((gsp->msgq.cnt - rptr) * GSP_PAGE_SIZE) - sizeof(*mqe); in r535_gsp_msgq_wait()
133 if (rptr == gsp->msgq.cnt) in r535_gsp_msgq_wait()
139 mqe = (void *)((u8 *)gsp->shm.msgq.ptr + 0x1000 + 0 * 0x1000); in r535_gsp_msgq_wait()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/sec2/
H A Dgp102.c127 ret = nvkm_falcon_msgq_recv_initmsg(sec2->msgq, &msg, sizeof(msg)); in gp102_sec2_initmsg()
137 nvkm_falcon_msgq_init(sec2->msgq, in gp102_sec2_initmsg()
172 if (!nvkm_falcon_msgq_empty(sec2->msgq)) in gp102_sec2_intr()
173 nvkm_falcon_msgq_recv(sec2->msgq); in gp102_sec2_intr()
215 .msgq = { 0xa30, 0xa34, 8 },
H A Dga102.c36 ret = nvkm_falcon_msgq_recv_initmsg(sec2->msgq, &msg, sizeof(msg)); in ga102_sec2_initmsg()
46 nvkm_falcon_msgq_init(sec2->msgq, msg.queue_info[i].index, in ga102_sec2_initmsg()
138 .msgq = { 0xc80, 0xc84, 8 },
H A Dbase.c118 nvkm_falcon_msgq_del(&sec2->msgq); in nvkm_sec2_dtor()
160 (ret = nvkm_falcon_msgq_new(sec2->qmgr, "msgq", &sec2->msgq))) in nvkm_sec2_new_()
H A Dtu102.c45 .msgq = { 0xc80, 0xc84, 8 },
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/
H A Dgm20b.c164 ret = nvkm_falcon_msgq_recv_initmsg(pmu->msgq, &msg, sizeof(msg)); in gm20b_pmu_initmsg()
178 nvkm_falcon_msgq_init(pmu->msgq, msg.queue_info[4].index, in gm20b_pmu_initmsg()
197 nvkm_falcon_msgq_recv(pmu->msgq); in gm20b_pmu_recv()
H A Dbase.c108 nvkm_falcon_msgq_del(&pmu->msgq); in nvkm_pmu_dtor()
152 (ret = nvkm_falcon_msgq_new(pmu->qmgr, "msgq", &pmu->msgq))) in nvkm_pmu_ctor()
H A Dgp102.c41 .msgq = { 0x4c8, 0x4cc, 0 },
H A Dgm200.c58 .msgq = { 0x4c8, 0x4cc, 0 },
/linux/drivers/isdn/mISDN/
H A Dstack.c28 skb_queue_tail(&st->msgq, skb); in _queue_message()
221 skb = skb_dequeue(&st->msgq); in mISDNStackd()
226 skb = skb_dequeue(&st->msgq); in mISDNStackd()
266 if (!skb_queue_empty(&st->msgq)) in mISDNStackd()
315 skb_queue_purge(&st->msgq); in mISDNStackd()
379 skb_queue_head_init(&newst->msgq); in create_stack()
/linux/drivers/gpu/drm/nouveau/include/nvkm/engine/
H A Dsec2.h18 struct nvkm_falcon_msgq *msgq; member
H A Dfalcon.h89 } cmdq, msgq; member
/linux/drivers/gpu/drm/nouveau/include/nvkm/subdev/
H A Dpmu.h15 struct nvkm_falcon_msgq *msgq; member
H A Dgsp.h121 } cmdq, msgq; member
144 } msgq; member
/linux/include/linux/
H A DmISDNif.h511 struct sk_buff_head msgq; member

12