/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_ih.c | 51 ih->ptr_mask = ih->ring_size - 1; in amdgpu_ih_ring_init() 158 wptr &= ih->ptr_mask; in amdgpu_ih_ring_write() 173 uint32_t cur_rptr = ih->rptr | (*prev_rptr & ~ih->ptr_mask); in amdgpu_ih_has_checkpoint_processed() 177 cur_rptr += ih->ptr_mask + 1; in amdgpu_ih_has_checkpoint_processed() 182 (cur_rptr & ih->ptr_mask) == amdgpu_ih_get_wptr(adev, ih); in amdgpu_ih_has_checkpoint_processed() 208 checkpoint_wptr += ih->ptr_mask + 1; in amdgpu_ih_wait_on_checkpoint_process() 243 ih->rptr &= ih->ptr_mask; in amdgpu_ih_process()
|
D | si_ih.c | 117 wptr, ih->rptr, (wptr + 16) & ih->ptr_mask); in si_ih_get_wptr() 118 ih->rptr = (wptr + 16) & ih->ptr_mask; in si_ih_get_wptr() 123 return (wptr & ih->ptr_mask); in si_ih_get_wptr()
|
D | cik_ih.c | 202 wptr, ih->rptr, (wptr + 16) & ih->ptr_mask); in cik_ih_get_wptr() 203 ih->rptr = (wptr + 16) & ih->ptr_mask; in cik_ih_get_wptr() 208 return (wptr & ih->ptr_mask); in cik_ih_get_wptr()
|
D | iceland_ih.c | 212 wptr, ih->rptr, (wptr + 16) & ih->ptr_mask); in iceland_ih_get_wptr() 213 ih->rptr = (wptr + 16) & ih->ptr_mask; in iceland_ih_get_wptr() 220 return (wptr & ih->ptr_mask); in iceland_ih_get_wptr()
|
D | cz_ih.c | 213 wptr, ih->rptr, (wptr + 16) & ih->ptr_mask); in cz_ih_get_wptr() 214 ih->rptr = (wptr + 16) & ih->ptr_mask; in cz_ih_get_wptr() 221 return (wptr & ih->ptr_mask); in cz_ih_get_wptr()
|
D | tonga_ih.c | 216 wptr, ih->rptr, (wptr + 16) & ih->ptr_mask); in tonga_ih_get_wptr() 217 ih->rptr = (wptr + 16) & ih->ptr_mask; in tonga_ih_get_wptr() 223 return (wptr & ih->ptr_mask); in tonga_ih_get_wptr()
|
D | amdgpu_ring.h | 222 uint64_t ptr_mask; member 317 ring->wptr &= ring->ptr_mask; in amdgpu_ring_write() 348 ring->wptr &= ring->ptr_mask; in amdgpu_ring_write_multiple()
|
D | amdgpu_ih.h | 50 uint32_t ptr_mask; member
|
D | vega10_ih.c | 364 tmp = (wptr + 32) & ih->ptr_mask; in vega10_ih_get_wptr() 375 return (wptr & ih->ptr_mask); in vega10_ih_get_wptr()
|
D | vega20_ih.c | 415 tmp = (wptr + 32) & ih->ptr_mask; in vega20_ih_get_wptr() 426 return (wptr & ih->ptr_mask); in vega20_ih_get_wptr()
|
D | navi10_ih.c | 441 tmp = (wptr + 32) & ih->ptr_mask; in navi10_ih_get_wptr() 451 return (wptr & ih->ptr_mask); in navi10_ih_get_wptr()
|
/drivers/net/ethernet/sfc/ |
D | tx_common.c | 18 return DIV_ROUND_UP(tx_queue->ptr_mask + 1, in efx_tx_cb_page_count() 31 tx_queue->ptr_mask = entries - 1; in efx_probe_tx_queue() 35 tx_queue->queue, efx->txq_entries, tx_queue->ptr_mask); in efx_probe_tx_queue() 113 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in efx_fini_tx_queue() 207 stop_index = (index + 1) & tx_queue->ptr_mask; in efx_dequeue_buffers() 208 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in efx_dequeue_buffers() 224 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in efx_dequeue_buffers() 246 EFX_WARN_ON_ONCE_PARANOID(index > tx_queue->ptr_mask); in efx_xmit_done()
|
D | ef100_tx.c | 27 (tx_queue->ptr_mask + 2) * in ef100_tx_probe() 120 write_ptr = tx_queue->write_count & tx_queue->ptr_mask; in ef100_notify_tx_desc() 280 write_ptr = new_write_count & tx_queue->ptr_mask; in ef100_tx_make_descriptors() 340 tx_queue->ptr_mask; in ef100_ev_tx() 362 if (!tx_queue->buffer || !tx_queue->ptr_mask) { in ef100_enqueue_skb()
|
D | ef100_rx.c | 145 rx_queue->removed_count & rx_queue->ptr_mask); in efx_ef100_ev_rx() 158 idx = rx_queue->notified_count & rx_queue->ptr_mask; in ef100_rx_write() 169 rx_queue->added_count & rx_queue->ptr_mask); in ef100_rx_write()
|
D | farch.c | 287 write_ptr = tx_queue->write_count & tx_queue->ptr_mask; in efx_farch_notify_tx_desc() 303 write_ptr = tx_queue->write_count & tx_queue->ptr_mask; in efx_farch_push_tx_desc() 328 write_ptr = tx_queue->write_count & tx_queue->ptr_mask; in efx_farch_tx_write() 349 old_write_count & tx_queue->ptr_mask); in efx_farch_tx_write() 377 entries = tx_queue->ptr_mask + 1; in efx_farch_tx_probe() 491 rx_queue->notified_count & rx_queue->ptr_mask); in efx_farch_rx_write() 496 write_ptr = rx_queue->added_count & rx_queue->ptr_mask; in efx_farch_rx_write() 507 entries = rx_queue->ptr_mask + 1; in efx_farch_rx_probe() 947 rx_queue->ptr_mask)) { in efx_farch_handle_rx_bad_index() 952 expected = rx_queue->removed_count & rx_queue->ptr_mask; in efx_farch_handle_rx_bad_index() [all …]
|
D | mcdi_functions.c | 178 MCDI_SET_DWORD(inbuf, INIT_TXQ_IN_SIZE, tx_queue->ptr_mask + 1); in efx_mcdi_tx_init() 273 (rx_queue->ptr_mask + 1) * in efx_mcdi_rx_probe() 297 MCDI_SET_DWORD(inbuf, INIT_RXQ_IN_SIZE, rx_queue->ptr_mask + 1); in efx_mcdi_rx_init()
|
D | net_driver.h | 271 unsigned int ptr_mask; member 395 unsigned int ptr_mask; member 1657 if (unlikely(rx_buf == efx_rx_buffer(rx_queue, rx_queue->ptr_mask))) in efx_rx_buf_next() 1737 return tx_queue->insert_count & tx_queue->ptr_mask; in efx_tx_queue_get_insert_index()
|
D | rx_common.c | 214 rx_queue->ptr_mask = entries - 1; in efx_probe_rx_queue() 219 rx_queue->ptr_mask); in efx_probe_rx_queue() 305 unsigned int index = i & rx_queue->ptr_mask; in efx_fini_rx_queue() 423 index = rx_queue->added_count & rx_queue->ptr_mask; in efx_init_rx_buffers()
|
D | tx.c | 564 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in efx_xmit_done_single() 585 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in efx_xmit_done_single()
|
D | ef10.c | 2213 (tx_queue->ptr_mask + 1) * in efx_ef10_tx_probe() 2225 write_ptr = tx_queue->write_count & tx_queue->ptr_mask; in efx_ef10_push_tx_desc() 2424 write_ptr = tx_queue->write_count & tx_queue->ptr_mask; in efx_ef10_notify_tx_desc() 2461 write_ptr = tx_queue->write_count & tx_queue->ptr_mask; in efx_ef10_tx_write() 2488 old_write_count & tx_queue->ptr_mask); in efx_ef10_tx_write() 2604 rx_queue->notified_count & rx_queue->ptr_mask); in efx_ef10_rx_write() 2609 write_count & rx_queue->ptr_mask); in efx_ef10_rx_write() 2677 unsigned int dropped = (actual - expected) & rx_queue->ptr_mask; in efx_ef10_handle_rx_bad_lbits() 2696 rx_desc_ptr = rx_queue->removed_count & rx_queue->ptr_mask; in efx_ef10_handle_rx_abort() 2927 rx_queue->removed_count & rx_queue->ptr_mask, in efx_ef10_handle_rx_event() [all …]
|
/drivers/gpu/drm/radeon/ |
D | radeon_ring.c | 88 ring->ring_free_dw &= ring->ptr_mask; in radeon_ring_free_size() 316 size &= ring->ptr_mask; in radeon_ring_backup() 330 ptr &= ring->ptr_mask; in radeon_ring_backup() 417 ring->ptr_mask = (ring->ring_size / 4) - 1; in radeon_ring_init() 506 i = (rptr + ring->ptr_mask + 1 - 32) & ring->ptr_mask; in radeon_debugfs_ring_info_show() 514 i = (i + 1) & ring->ptr_mask; in radeon_debugfs_ring_info_show()
|
/drivers/net/ethernet/sfc/falcon/ |
D | tx.c | 362 stop_index = (index + 1) & tx_queue->ptr_mask; in ef4_dequeue_buffers() 363 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in ef4_dequeue_buffers() 380 read_ptr = tx_queue->read_count & tx_queue->ptr_mask; in ef4_dequeue_buffers() 498 EF4_BUG_ON_PARANOID(index > tx_queue->ptr_mask); in ef4_xmit_done() 535 return DIV_ROUND_UP(tx_queue->ptr_mask + 1, PAGE_SIZE >> EF4_TX_CB_ORDER); in ef4_tx_cb_page_count() 547 tx_queue->ptr_mask = entries - 1; in ef4_probe_tx_queue() 551 tx_queue->queue, efx->txq_entries, tx_queue->ptr_mask); in ef4_probe_tx_queue() 619 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in ef4_fini_tx_queue()
|
D | farch.c | 277 write_ptr = tx_queue->write_count & tx_queue->ptr_mask; in ef4_farch_notify_tx_desc() 293 write_ptr = tx_queue->write_count & tx_queue->ptr_mask; in ef4_farch_push_tx_desc() 318 write_ptr = tx_queue->write_count & tx_queue->ptr_mask; in ef4_farch_tx_write() 339 old_write_count & tx_queue->ptr_mask); in ef4_farch_tx_write() 368 entries = tx_queue->ptr_mask + 1; in ef4_farch_tx_probe() 498 rx_queue->notified_count & rx_queue->ptr_mask); in ef4_farch_rx_write() 503 write_ptr = rx_queue->added_count & rx_queue->ptr_mask; in ef4_farch_rx_write() 514 entries = rx_queue->ptr_mask + 1; in ef4_farch_rx_probe() 844 tx_queue->ptr_mask); in ef4_farch_handle_tx_event() 955 rx_queue->ptr_mask)) { in ef4_farch_handle_rx_bad_index() [all …]
|
D | rx.c | 78 if (unlikely(rx_buf == ef4_rx_buffer(rx_queue, rx_queue->ptr_mask))) in ef4_rx_buf_next() 191 index = rx_queue->added_count & rx_queue->ptr_mask; in ef4_init_rx_buffers() 560 (index + n_frags - 1) & rx_queue->ptr_mask, len, in ef4_rx_packet() 695 rx_queue->ptr_mask = entries - 1; in ef4_probe_rx_queue() 700 rx_queue->ptr_mask); in ef4_probe_rx_queue() 798 unsigned index = i & rx_queue->ptr_mask; in ef4_fini_rx_queue()
|
D | net_driver.h | 221 unsigned int ptr_mask; member 333 unsigned int ptr_mask; member 1312 return tx_queue->insert_count & tx_queue->ptr_mask; in ef4_tx_queue_get_insert_index()
|