Lines Matching full:ring
11 struct ath12k_dbring *ring, in ath12k_dbring_bufs_replenish() argument
23 srng = &ab->hal.srng_list[ring->refill_srng.ring_id]; in ath12k_dbring_bufs_replenish()
30 ptr_aligned = PTR_ALIGN(ptr_unaligned, ring->buf_align); in ath12k_dbring_bufs_replenish()
31 paddr = dma_map_single(ab->dev, ptr_aligned, ring->buf_sz, in ath12k_dbring_bufs_replenish()
38 spin_lock_bh(&ring->idr_lock); in ath12k_dbring_bufs_replenish()
39 buf_id = idr_alloc(&ring->bufs_idr, buff, 0, ring->bufs_max, gfp); in ath12k_dbring_bufs_replenish()
40 spin_unlock_bh(&ring->idr_lock); in ath12k_dbring_bufs_replenish()
64 spin_lock_bh(&ring->idr_lock); in ath12k_dbring_bufs_replenish()
65 idr_remove(&ring->bufs_idr, buf_id); in ath12k_dbring_bufs_replenish()
66 spin_unlock_bh(&ring->idr_lock); in ath12k_dbring_bufs_replenish()
68 dma_unmap_single(ab->dev, paddr, ring->buf_sz, in ath12k_dbring_bufs_replenish()
76 struct ath12k_dbring *ring, in ath12k_dbring_fill_bufs() argument
86 srng = &ab->hal.srng_list[ring->refill_srng.ring_id]; in ath12k_dbring_fill_bufs()
91 req_entries = min(num_free, ring->bufs_max); in ath12k_dbring_fill_bufs()
93 align = ring->buf_align; in ath12k_dbring_fill_bufs()
94 size = sizeof(*buff) + ring->buf_sz + align - 1; in ath12k_dbring_fill_bufs()
101 ret = ath12k_dbring_bufs_replenish(ar, ring, buff, gfp); in ath12k_dbring_fill_bufs()
103 ath12k_warn(ab, "failed to replenish db ring num_remain %d req_ent %d\n", in ath12k_dbring_fill_bufs()
117 struct ath12k_dbring *ring, in ath12k_dbring_wmi_cfg_setup() argument
126 arg.pdev_id = DP_SW2HW_MACID(ring->pdev_id); in ath12k_dbring_wmi_cfg_setup()
128 arg.base_paddr_lo = lower_32_bits(ring->refill_srng.paddr); in ath12k_dbring_wmi_cfg_setup()
129 arg.base_paddr_hi = upper_32_bits(ring->refill_srng.paddr); in ath12k_dbring_wmi_cfg_setup()
130 arg.head_idx_paddr_lo = lower_32_bits(ring->hp_addr); in ath12k_dbring_wmi_cfg_setup()
131 arg.head_idx_paddr_hi = upper_32_bits(ring->hp_addr); in ath12k_dbring_wmi_cfg_setup()
132 arg.tail_idx_paddr_lo = lower_32_bits(ring->tp_addr); in ath12k_dbring_wmi_cfg_setup()
133 arg.tail_idx_paddr_hi = upper_32_bits(ring->tp_addr); in ath12k_dbring_wmi_cfg_setup()
134 arg.num_elems = ring->bufs_max; in ath12k_dbring_wmi_cfg_setup()
135 arg.buf_size = ring->buf_sz; in ath12k_dbring_wmi_cfg_setup()
136 arg.num_resp_per_event = ring->num_resp_per_event; in ath12k_dbring_wmi_cfg_setup()
137 arg.event_timeout_ms = ring->event_timeout_ms; in ath12k_dbring_wmi_cfg_setup()
141 ath12k_warn(ar->ab, "failed to setup db ring cfg\n"); in ath12k_dbring_wmi_cfg_setup()
148 int ath12k_dbring_set_cfg(struct ath12k *ar, struct ath12k_dbring *ring, in ath12k_dbring_set_cfg() argument
153 if (WARN_ON(!ring)) in ath12k_dbring_set_cfg()
156 ring->num_resp_per_event = num_resp_per_event; in ath12k_dbring_set_cfg()
157 ring->event_timeout_ms = event_timeout_ms; in ath12k_dbring_set_cfg()
158 ring->handler = handler; in ath12k_dbring_set_cfg()
164 struct ath12k_dbring *ring, in ath12k_dbring_buf_setup() argument
171 srng = &ab->hal.srng_list[ring->refill_srng.ring_id]; in ath12k_dbring_buf_setup()
172 ring->bufs_max = ring->refill_srng.size / in ath12k_dbring_buf_setup()
175 ring->buf_sz = db_cap->min_buf_sz; in ath12k_dbring_buf_setup()
176 ring->buf_align = db_cap->min_buf_align; in ath12k_dbring_buf_setup()
177 ring->pdev_id = db_cap->pdev_id; in ath12k_dbring_buf_setup()
178 ring->hp_addr = ath12k_hal_srng_get_hp_addr(ab, srng); in ath12k_dbring_buf_setup()
179 ring->tp_addr = ath12k_hal_srng_get_tp_addr(ab, srng); in ath12k_dbring_buf_setup()
181 ret = ath12k_dbring_fill_bufs(ar, ring, GFP_KERNEL); in ath12k_dbring_buf_setup()
186 int ath12k_dbring_srng_setup(struct ath12k *ar, struct ath12k_dbring *ring, in ath12k_dbring_srng_setup() argument
191 ret = ath12k_dp_srng_setup(ar->ab, &ring->refill_srng, HAL_RXDMA_DIR_BUF, in ath12k_dbring_srng_setup()
201 ath12k_dp_srng_cleanup(ar->ab, &ring->refill_srng); in ath12k_dbring_srng_setup()
233 struct ath12k_dbring *ring = NULL; in ath12k_dbring_buffer_release_event() local
275 ring = NULL; in ath12k_dbring_buffer_release_event()
281 if (!ring) { in ath12k_dbring_buffer_release_event()
286 srng = &ab->hal.srng_list[ring->refill_srng.ring_id]; in ath12k_dbring_buffer_release_event()
288 size = sizeof(*buff) + ring->buf_sz + ring->buf_align - 1; in ath12k_dbring_buffer_release_event()
304 spin_lock_bh(&ring->idr_lock); in ath12k_dbring_buffer_release_event()
305 buff = idr_find(&ring->bufs_idr, buf_id); in ath12k_dbring_buffer_release_event()
307 spin_unlock_bh(&ring->idr_lock); in ath12k_dbring_buffer_release_event()
310 idr_remove(&ring->bufs_idr, buf_id); in ath12k_dbring_buffer_release_event()
311 spin_unlock_bh(&ring->idr_lock); in ath12k_dbring_buffer_release_event()
313 dma_unmap_single(ab->dev, buff->paddr, ring->buf_sz, in ath12k_dbring_buffer_release_event()
316 if (ring->handler) { in ath12k_dbring_buffer_release_event()
319 ring->buf_align); in ath12k_dbring_buffer_release_event()
320 handler_data.data_sz = ring->buf_sz; in ath12k_dbring_buffer_release_event()
322 ring->handler(ar, &handler_data); in ath12k_dbring_buffer_release_event()
326 ath12k_dbring_bufs_replenish(ar, ring, buff, GFP_ATOMIC); in ath12k_dbring_buffer_release_event()
337 void ath12k_dbring_srng_cleanup(struct ath12k *ar, struct ath12k_dbring *ring) in ath12k_dbring_srng_cleanup() argument
339 ath12k_dp_srng_cleanup(ar->ab, &ring->refill_srng); in ath12k_dbring_srng_cleanup()
342 void ath12k_dbring_buf_cleanup(struct ath12k *ar, struct ath12k_dbring *ring) in ath12k_dbring_buf_cleanup() argument
347 spin_lock_bh(&ring->idr_lock); in ath12k_dbring_buf_cleanup()
348 idr_for_each_entry(&ring->bufs_idr, buff, buf_id) { in ath12k_dbring_buf_cleanup()
349 idr_remove(&ring->bufs_idr, buf_id); in ath12k_dbring_buf_cleanup()
351 ring->buf_sz, DMA_FROM_DEVICE); in ath12k_dbring_buf_cleanup()
355 idr_destroy(&ring->bufs_idr); in ath12k_dbring_buf_cleanup()
356 spin_unlock_bh(&ring->idr_lock); in ath12k_dbring_buf_cleanup()