/drivers/gpu/drm/ |
D | drm_bufs.c | 579 if (entry->buf_count) { in drm_cleanup_buf_error() 580 for (i = 0; i < entry->buf_count; i++) { in drm_cleanup_buf_error() 585 entry->buf_count = 0; in drm_cleanup_buf_error() 668 if (entry->buf_count) { in drm_addbufs_agp() 692 while (entry->buf_count < count) { in drm_addbufs_agp() 693 buf = &entry->buflist[entry->buf_count]; in drm_addbufs_agp() 694 buf->idx = dma->buf_count + entry->buf_count; in drm_addbufs_agp() 711 entry->buf_count = count; in drm_addbufs_agp() 718 DRM_DEBUG("buffer %d @ %p\n", entry->buf_count, buf->address); in drm_addbufs_agp() 721 entry->buf_count++; in drm_addbufs_agp() [all …]
|
D | drm_dma.c | 83 dma->bufs[i].buf_count, in drm_dma_takedown() 92 if (dma->bufs[i].buf_count) { in drm_dma_takedown() 93 for (j = 0; j < dma->bufs[i].buf_count; j++) { in drm_dma_takedown() 140 for (i = 0; i < dma->buf_count; i++) { in drm_core_reclaim_buffers()
|
D | drm_info.c | 131 if (dma->bufs[i].buf_count) { in drm_bufs_info() 136 dma->bufs[i].buf_count, in drm_bufs_info() 144 for (i = 0; i < dma->buf_count; i++) { in drm_bufs_info()
|
/drivers/staging/ozwpan/ |
D | ozeltbuf.c | 177 st->buf_count -= ei->length; in oz_elt_stream_delete() 179 st->buf_count, in oz_elt_stream_delete() 242 st->buf_count += ei->length; in oz_queue_elt_info() 246 st->buf_count, ei->length); in oz_queue_elt_info() 250 while ((st->buf_count > st->max_buf_count) && in oz_queue_elt_info() 257 st->buf_count -= ei2->length; in oz_queue_elt_info() 295 ei->stream->buf_count -= ei->length; in oz_select_elts_for_tx() 298 ei->stream->buf_count, ei->length); in oz_select_elts_for_tx()
|
D | ozeltbuf.h | 20 unsigned buf_count; member
|
/drivers/net/ethernet/sfc/ |
D | siena_sriov.c | 502 static bool bad_buf_count(unsigned buf_count, unsigned max_entry_count) in bad_buf_count() argument 507 return ((buf_count & (buf_count - 1)) || buf_count > max_buf_count); in bad_buf_count() 536 unsigned buf_count = req->u.init_evq.buf_count; in efx_vfdi_init_evq() local 542 bad_buf_count(buf_count, EFX_MAX_VF_EVQ_SIZE)) { in efx_vfdi_init_evq() 546 vf->pci_name, vf_evq, buf_count); in efx_vfdi_init_evq() 550 efx_sriov_bufs(efx, buftbl, req->u.init_evq.addr, buf_count); in efx_vfdi_init_evq() 559 FRF_AZ_EVQ_SIZE, __ffs(buf_count), in efx_vfdi_init_evq() 565 buf_count * sizeof(u64)); in efx_vfdi_init_evq() 566 vf->evq0_count = buf_count; in efx_vfdi_init_evq() 578 unsigned buf_count = req->u.init_rxq.buf_count; in efx_vfdi_init_rxq() local [all …]
|
D | vfdi.h | 173 u32 buf_count; member 178 u32 buf_count; member 188 u32 buf_count; member
|
/drivers/media/platform/s5p-mfc/ |
D | s5p_mfc_dec.c | 903 const struct v4l2_format *fmt, unsigned int *buf_count, in s5p_mfc_queue_setup() argument 916 if (*buf_count < 1) in s5p_mfc_queue_setup() 917 *buf_count = 1; in s5p_mfc_queue_setup() 918 if (*buf_count > MFC_MAX_BUFFERS) in s5p_mfc_queue_setup() 919 *buf_count = MFC_MAX_BUFFERS; in s5p_mfc_queue_setup() 927 if (*buf_count < ctx->pb_count) in s5p_mfc_queue_setup() 928 *buf_count = ctx->pb_count; in s5p_mfc_queue_setup() 929 if (*buf_count > ctx->pb_count + MFC_MAX_EXTRA_DPB) in s5p_mfc_queue_setup() 930 *buf_count = ctx->pb_count + MFC_MAX_EXTRA_DPB; in s5p_mfc_queue_setup() 931 if (*buf_count > MFC_MAX_BUFFERS) in s5p_mfc_queue_setup() [all …]
|
D | s5p_mfc_enc.c | 1634 unsigned int *buf_count, unsigned int *plane_count, in s5p_mfc_queue_setup() argument 1649 if (*buf_count < 1) in s5p_mfc_queue_setup() 1650 *buf_count = 1; in s5p_mfc_queue_setup() 1651 if (*buf_count > MFC_MAX_BUFFERS) in s5p_mfc_queue_setup() 1652 *buf_count = MFC_MAX_BUFFERS; in s5p_mfc_queue_setup() 1661 if (*buf_count < 1) in s5p_mfc_queue_setup() 1662 *buf_count = 1; in s5p_mfc_queue_setup() 1663 if (*buf_count > MFC_MAX_BUFFERS) in s5p_mfc_queue_setup() 1664 *buf_count = MFC_MAX_BUFFERS; in s5p_mfc_queue_setup()
|
/drivers/mtd/nand/ |
D | pxa3xx_nand.c | 158 unsigned int buf_count; member 506 info->buf_count = 0; in prepare_command_pool() 571 info->buf_count = mtd->writesize + mtd->oobsize; in prepare_command_pool() 572 memset(info->data_buff, 0xFF, info->buf_count); in prepare_command_pool() 594 info->buf_count = host->read_id_bytes; in prepare_command_pool() 603 info->buf_count = 1; in prepare_command_pool() 735 if (info->buf_start < info->buf_count) in pxa3xx_nand_read_byte() 748 if (!(info->buf_start & 0x01) && info->buf_start < info->buf_count) { in pxa3xx_nand_read_word() 759 int real_len = min_t(size_t, len, info->buf_count - info->buf_start); in pxa3xx_nand_read_buf() 770 int real_len = min_t(size_t, len, info->buf_count - info->buf_start); in pxa3xx_nand_write_buf()
|
/drivers/gpu/drm/r128/ |
D | r128_cce.c | 780 for (i = 0; i < dma->buf_count; i++) { 820 for (i = 0; i < dma->buf_count; i++) { in r128_freelist_get() 830 for (i = 0; i < dma->buf_count; i++) { in r128_freelist_get() 853 for (i = 0; i < dma->buf_count; i++) { in r128_freelist_reset() 925 if (d->request_count < 0 || d->request_count > dma->buf_count) { in r128_cce_buffers() 927 DRM_CURRENTPID, d->request_count, dma->buf_count); in r128_cce_buffers()
|
D | r128_state.c | 1355 if (vertex->idx < 0 || vertex->idx >= dma->buf_count) { in r128_cce_vertex() 1357 vertex->idx, dma->buf_count - 1); in r128_cce_vertex() 1408 if (elts->idx < 0 || elts->idx >= dma->buf_count) { in r128_cce_indices() 1410 elts->idx, dma->buf_count - 1); in r128_cce_indices() 1470 if (blit->idx < 0 || blit->idx >= dma->buf_count) { in r128_cce_blit() 1472 blit->idx, dma->buf_count - 1); in r128_cce_blit() 1557 if (indirect->idx < 0 || indirect->idx >= dma->buf_count) { in r128_cce_indirect() 1559 indirect->idx, dma->buf_count - 1); in r128_cce_indirect()
|
/drivers/gpu/drm/i810/ |
D | i810_dma.c | 56 for (i = 0; i < dma->buf_count; i++) { in i810_freelist_get() 226 for (i = 0; i < dma->buf_count; i++) { in i810_dma_cleanup() 289 if (dma->buf_count > 1019) { in i810_freelist_init() 294 for (i = 0; i < dma->buf_count; i++) { in i810_freelist_init() 869 for (i = 0; i < dma->buf_count; i++) { in i810_flush_queue() 901 for (i = 0; i < dma->buf_count; i++) { in i810_driver_reclaim_buffers() 941 if (vertex->idx < 0 || vertex->idx > dma->buf_count) in i810_dma_vertex() 1103 if (mc->idx >= dma->buf_count || mc->idx < 0) in i810_dma_mc()
|
/drivers/usb/host/ |
D | isp1362-hcd.c | 205 found = bitmap_find_next_zero_area(&epq->buf_map, epq->buf_count, 0, in claim_ptd_buffers() 207 if (found >= epq->buf_count) in claim_ptd_buffers() 217 BUG_ON(epq->buf_avail > epq->buf_count); in claim_ptd_buffers() 222 epq->buf_avail, epq->buf_count, num_ptds, epq->buf_map, epq->skip_map); in claim_ptd_buffers() 231 if (last > epq->buf_count) in release_ptd_buffers() 234 ep->ptd_offset, ep->num_ptds, epq->buf_count, epq->buf_avail, in release_ptd_buffers() 236 BUG_ON(last > epq->buf_count); in release_ptd_buffers() 243 BUG_ON(epq->buf_avail > epq->buf_count); in release_ptd_buffers() 244 BUG_ON(epq->ptd_count > epq->buf_count); in release_ptd_buffers() 248 ep->ptd_offset, ep->num_ptds, epq->buf_avail, epq->buf_count); in release_ptd_buffers() [all …]
|
/drivers/gpu/drm/mga/ |
D | mga_dma.c | 251 DRM_DEBUG("count=%d\n", dma->buf_count); in mga_freelist_init() 259 for (i = 0; i < dma->buf_count; i++) { in mga_freelist_init() 314 for (i = 0; i < dma->buf_count; i++) { 1115 if (d->request_count < 0 || d->request_count > dma->buf_count) { in mga_dma_buffers() 1117 DRM_CURRENTPID, d->request_count, dma->buf_count); in mga_dma_buffers()
|
D | mga_state.c | 882 if (vertex->idx < 0 || vertex->idx > dma->buf_count) in mga_dma_vertex() 917 if (indices->idx < 0 || indices->idx > dma->buf_count) in mga_dma_indices() 960 if (iload->idx < 0 || iload->idx > dma->buf_count) in mga_dma_iload()
|
/drivers/media/rc/ |
D | fintek-cir.h | 73 unsigned int buf_count; member
|
D | nuvoton-cir.h | 80 unsigned int buf_count; member
|
D | nuvoton-cir.c | 563 nvt->tx.buf_count = (ret * sizeof(unsigned)); in nvt_tx_ir() 565 memcpy(nvt->tx.buf, txbuf, nvt->tx.buf_count); in nvt_tx_ir() 835 count = nvt->tx.buf_count; in nvt_cir_isr()
|
/drivers/gpu/drm/savage/ |
D | savage_bci.c | 213 DRM_DEBUG("count=%d\n", dma->buf_count); in savage_freelist_init() 223 for (i = 0; i < dma->buf_count; i++) { in savage_freelist_init() 1036 if (d->request_count < 0 || d->request_count > dma->buf_count) { in savage_bci_buffers() 1038 DRM_CURRENTPID, d->request_count, dma->buf_count); in savage_bci_buffers() 1070 for (i = 0; i < dma->buf_count; i++) { in savage_reclaim_buffers()
|
/drivers/net/ethernet/pasemi/ |
D | pasemi_mac.c | 870 unsigned int start, descr_count, buf_count, batch_limit; in pasemi_mac_clean_tx() local 893 buf_count = 0; in pasemi_mac_clean_tx() 898 i += buf_count) { in pasemi_mac_clean_tx() 910 buf_count = 2; in pasemi_mac_clean_tx() 921 buf_count = 2 + nr_frags; in pasemi_mac_clean_tx() 925 if (buf_count & 1) in pasemi_mac_clean_tx() 926 buf_count++; in pasemi_mac_clean_tx()
|
/drivers/s390/net/ |
D | qeth_core.h | 235 ((card)->qdio.in_buf_pool.buf_count / 2) 238 #define QETH_PCI_THRESHOLD_A(card) ((card)->qdio.in_buf_pool.buf_count+1) 412 int buf_count; member
|
D | qeth_core_sys.c | 290 return sprintf(buf, "%i\n", card->qdio.in_buf_pool.buf_count); in qeth_dev_bufcnt_show() 311 old_cnt = card->qdio.in_buf_pool.buf_count; in qeth_dev_bufcnt_store()
|
/drivers/gpu/drm/radeon/ |
D | radeon_state.c | 2272 if (vertex->idx < 0 || vertex->idx >= dma->buf_count) { in radeon_cp_vertex() 2274 vertex->idx, dma->buf_count - 1); in radeon_cp_vertex() 2353 if (elts->idx < 0 || elts->idx >= dma->buf_count) { in radeon_cp_indices() 2355 elts->idx, dma->buf_count - 1); in radeon_cp_indices() 2488 if (indirect->idx < 0 || indirect->idx >= dma->buf_count) { in radeon_cp_indirect() 2490 indirect->idx, dma->buf_count - 1); in radeon_cp_indirect() 2559 if (vertex->idx < 0 || vertex->idx >= dma->buf_count) { in radeon_cp_vertex2() 2561 vertex->idx, dma->buf_count - 1); in radeon_cp_vertex2() 2953 if (idx < 0 || idx >= dma->buf_count) { in radeon_cp_cmdbuf() 2955 idx, dma->buf_count - 1); in radeon_cp_cmdbuf()
|
D | radeon_cp.c | 1928 if (++dev_priv->last_buf >= dma->buf_count) in radeon_freelist_get() 1936 for (i = 0; i < dma->buf_count; i++) { in radeon_freelist_get() 1946 if (++start >= dma->buf_count) in radeon_freelist_get() 1966 for (i = 0; i < dma->buf_count; i++) { in radeon_freelist_reset() 2053 if (d->request_count < 0 || d->request_count > dma->buf_count) { in radeon_cp_buffers() 2055 DRM_CURRENTPID, d->request_count, dma->buf_count); in radeon_cp_buffers()
|