Lines Matching full:ring
12 struct ath12k_dbring *ring, in ath12k_dbring_bufs_replenish() argument
24 srng = &ab->hal.srng_list[ring->refill_srng.ring_id]; in ath12k_dbring_bufs_replenish()
31 ptr_aligned = PTR_ALIGN(ptr_unaligned, ring->buf_align); in ath12k_dbring_bufs_replenish()
32 paddr = dma_map_single(ab->dev, ptr_aligned, ring->buf_sz, in ath12k_dbring_bufs_replenish()
39 spin_lock_bh(&ring->idr_lock); in ath12k_dbring_bufs_replenish()
40 buf_id = idr_alloc(&ring->bufs_idr, buff, 0, ring->bufs_max, gfp); in ath12k_dbring_bufs_replenish()
41 spin_unlock_bh(&ring->idr_lock); in ath12k_dbring_bufs_replenish()
65 spin_lock_bh(&ring->idr_lock); in ath12k_dbring_bufs_replenish()
66 idr_remove(&ring->bufs_idr, buf_id); in ath12k_dbring_bufs_replenish()
67 spin_unlock_bh(&ring->idr_lock); in ath12k_dbring_bufs_replenish()
69 dma_unmap_single(ab->dev, paddr, ring->buf_sz, in ath12k_dbring_bufs_replenish()
77 struct ath12k_dbring *ring, in ath12k_dbring_fill_bufs() argument
87 srng = &ab->hal.srng_list[ring->refill_srng.ring_id]; in ath12k_dbring_fill_bufs()
92 req_entries = min(num_free, ring->bufs_max); in ath12k_dbring_fill_bufs()
94 align = ring->buf_align; in ath12k_dbring_fill_bufs()
95 size = sizeof(*buff) + ring->buf_sz + align - 1; in ath12k_dbring_fill_bufs()
102 ret = ath12k_dbring_bufs_replenish(ar, ring, buff, gfp); in ath12k_dbring_fill_bufs()
104 ath12k_warn(ab, "failed to replenish db ring num_remain %d req_ent %d\n", in ath12k_dbring_fill_bufs()
118 struct ath12k_dbring *ring, in ath12k_dbring_wmi_cfg_setup() argument
127 arg.pdev_id = DP_SW2HW_MACID(ring->pdev_id); in ath12k_dbring_wmi_cfg_setup()
129 arg.base_paddr_lo = lower_32_bits(ring->refill_srng.paddr); in ath12k_dbring_wmi_cfg_setup()
130 arg.base_paddr_hi = upper_32_bits(ring->refill_srng.paddr); in ath12k_dbring_wmi_cfg_setup()
131 arg.head_idx_paddr_lo = lower_32_bits(ring->hp_addr); in ath12k_dbring_wmi_cfg_setup()
132 arg.head_idx_paddr_hi = upper_32_bits(ring->hp_addr); in ath12k_dbring_wmi_cfg_setup()
133 arg.tail_idx_paddr_lo = lower_32_bits(ring->tp_addr); in ath12k_dbring_wmi_cfg_setup()
134 arg.tail_idx_paddr_hi = upper_32_bits(ring->tp_addr); in ath12k_dbring_wmi_cfg_setup()
135 arg.num_elems = ring->bufs_max; in ath12k_dbring_wmi_cfg_setup()
136 arg.buf_size = ring->buf_sz; in ath12k_dbring_wmi_cfg_setup()
137 arg.num_resp_per_event = ring->num_resp_per_event; in ath12k_dbring_wmi_cfg_setup()
138 arg.event_timeout_ms = ring->event_timeout_ms; in ath12k_dbring_wmi_cfg_setup()
142 ath12k_warn(ar->ab, "failed to setup db ring cfg\n"); in ath12k_dbring_wmi_cfg_setup()
149 int ath12k_dbring_set_cfg(struct ath12k *ar, struct ath12k_dbring *ring, in ath12k_dbring_set_cfg() argument
154 if (WARN_ON(!ring)) in ath12k_dbring_set_cfg()
157 ring->num_resp_per_event = num_resp_per_event; in ath12k_dbring_set_cfg()
158 ring->event_timeout_ms = event_timeout_ms; in ath12k_dbring_set_cfg()
159 ring->handler = handler; in ath12k_dbring_set_cfg()
165 struct ath12k_dbring *ring, in ath12k_dbring_buf_setup() argument
172 srng = &ab->hal.srng_list[ring->refill_srng.ring_id]; in ath12k_dbring_buf_setup()
173 ring->bufs_max = ring->refill_srng.size / in ath12k_dbring_buf_setup()
176 ring->buf_sz = db_cap->min_buf_sz; in ath12k_dbring_buf_setup()
177 ring->buf_align = db_cap->min_buf_align; in ath12k_dbring_buf_setup()
178 ring->pdev_id = db_cap->pdev_id; in ath12k_dbring_buf_setup()
179 ring->hp_addr = ath12k_hal_srng_get_hp_addr(ab, srng); in ath12k_dbring_buf_setup()
180 ring->tp_addr = ath12k_hal_srng_get_tp_addr(ab, srng); in ath12k_dbring_buf_setup()
182 ret = ath12k_dbring_fill_bufs(ar, ring, GFP_KERNEL); in ath12k_dbring_buf_setup()
187 int ath12k_dbring_srng_setup(struct ath12k *ar, struct ath12k_dbring *ring, in ath12k_dbring_srng_setup() argument
192 ret = ath12k_dp_srng_setup(ar->ab, &ring->refill_srng, HAL_RXDMA_DIR_BUF, in ath12k_dbring_srng_setup()
202 ath12k_dp_srng_cleanup(ar->ab, &ring->refill_srng); in ath12k_dbring_srng_setup()
234 struct ath12k_dbring *ring = NULL; in ath12k_dbring_buffer_release_event() local
276 ring = NULL; in ath12k_dbring_buffer_release_event()
282 if (!ring) { in ath12k_dbring_buffer_release_event()
287 srng = &ab->hal.srng_list[ring->refill_srng.ring_id]; in ath12k_dbring_buffer_release_event()
289 size = sizeof(*buff) + ring->buf_sz + ring->buf_align - 1; in ath12k_dbring_buffer_release_event()
305 spin_lock_bh(&ring->idr_lock); in ath12k_dbring_buffer_release_event()
306 buff = idr_find(&ring->bufs_idr, buf_id); in ath12k_dbring_buffer_release_event()
308 spin_unlock_bh(&ring->idr_lock); in ath12k_dbring_buffer_release_event()
311 idr_remove(&ring->bufs_idr, buf_id); in ath12k_dbring_buffer_release_event()
312 spin_unlock_bh(&ring->idr_lock); in ath12k_dbring_buffer_release_event()
314 dma_unmap_single(ab->dev, buff->paddr, ring->buf_sz, in ath12k_dbring_buffer_release_event()
317 if (ring->handler) { in ath12k_dbring_buffer_release_event()
320 ring->buf_align); in ath12k_dbring_buffer_release_event()
321 handler_data.data_sz = ring->buf_sz; in ath12k_dbring_buffer_release_event()
323 ring->handler(ar, &handler_data); in ath12k_dbring_buffer_release_event()
327 ath12k_dbring_bufs_replenish(ar, ring, buff, GFP_ATOMIC); in ath12k_dbring_buffer_release_event()
338 void ath12k_dbring_srng_cleanup(struct ath12k *ar, struct ath12k_dbring *ring) in ath12k_dbring_srng_cleanup() argument
340 ath12k_dp_srng_cleanup(ar->ab, &ring->refill_srng); in ath12k_dbring_srng_cleanup()
343 void ath12k_dbring_buf_cleanup(struct ath12k *ar, struct ath12k_dbring *ring) in ath12k_dbring_buf_cleanup() argument
348 spin_lock_bh(&ring->idr_lock); in ath12k_dbring_buf_cleanup()
349 idr_for_each_entry(&ring->bufs_idr, buff, buf_id) { in ath12k_dbring_buf_cleanup()
350 idr_remove(&ring->bufs_idr, buf_id); in ath12k_dbring_buf_cleanup()
352 ring->buf_sz, DMA_FROM_DEVICE); in ath12k_dbring_buf_cleanup()
356 idr_destroy(&ring->bufs_idr); in ath12k_dbring_buf_cleanup()
357 spin_unlock_bh(&ring->idr_lock); in ath12k_dbring_buf_cleanup()