/drivers/crypto/ccree/ |
D | cc_hash.h | 49 u32 buf_cnt[2]; member 61 return &state->buf_cnt[state->buff_index]; in cc_hash_buf_cnt() 71 return &state->buf_cnt[state->buff_index ^ 1]; in cc_next_buf_cnt()
|
D | cc_hash.c | 1552 state->buf_cnt[0] = tmp; in cc_hash_import()
|
/drivers/net/ethernet/cavium/liquidio/ |
D | octeon_droq.c | 335 u32 buf_cnt, in octeon_create_recv_info() argument 353 recv_pkt->buffer_count = (u16)buf_cnt; in octeon_create_recv_info() 359 while (buf_cnt) { in octeon_create_recv_info() 379 buf_cnt--; in octeon_create_recv_info() 568 u32 i = 0, buf_cnt; in octeon_droq_drop_packets() local 579 buf_cnt = octeon_droq_get_bufcount(droq->buffer_size, in octeon_droq_drop_packets() 583 buf_cnt = 1; in octeon_droq_drop_packets() 586 droq->read_idx = incr_index(droq->read_idx, buf_cnt, in octeon_droq_drop_packets() 588 droq->refill_count += buf_cnt; in octeon_droq_drop_packets() 630 u32 buf_cnt; in octeon_droq_fast_process_packets() local [all …]
|
/drivers/infiniband/hw/hns/ |
D | hns_roce_alloc.c | 134 int buf_cnt, struct hns_roce_buf *buf, in hns_roce_get_kmem_bufs() argument 149 for (i = 0; i < buf_cnt && offset < max_size; i++) { in hns_roce_get_kmem_bufs() 158 int buf_cnt, struct ib_umem *umem, in hns_roce_get_umem_bufs() argument 167 if (total >= buf_cnt) in hns_roce_get_umem_bufs()
|
D | hns_roce_device.h | 1207 int buf_cnt, struct hns_roce_buf *buf, 1210 int buf_cnt, struct ib_umem *umem,
|
/drivers/crypto/keembay/ |
D | keembay-ocs-hcu-core.c | 87 size_t buf_cnt; member 117 return rctx->sg_data_total + rctx->buf_cnt; in kmb_get_total_data() 125 if (rctx->sg_data_total > (sizeof(rctx->buffer) - rctx->buf_cnt)) { in flush_sg_to_ocs_buffer() 152 scatterwalk_map_and_copy(&rctx->buffer[rctx->buf_cnt], in flush_sg_to_ocs_buffer() 158 rctx->buf_cnt += count; in flush_sg_to_ocs_buffer() 274 if (rctx->buf_cnt) { in kmb_ocs_dma_prepare() 276 rctx->buf_cnt, in kmb_ocs_dma_prepare() 283 rctx->buf_dma_count = rctx->buf_cnt; in kmb_ocs_dma_prepare() 382 WARN(rctx->buf_cnt, "%s: Context buffer is not empty\n", __func__); in prepare_ipad() 406 rctx->buf_cnt = rctx->blk_sz; in prepare_ipad() [all …]
|
/drivers/crypto/ |
D | sahara.c | 181 unsigned int buf_cnt; member 901 len = rctx->buf_cnt + req->nbytes; in sahara_sha_prepare_request() 906 scatterwalk_map_and_copy(rctx->buf + rctx->buf_cnt, req->src, in sahara_sha_prepare_request() 908 rctx->buf_cnt += req->nbytes; in sahara_sha_prepare_request() 914 if (rctx->buf_cnt) in sahara_sha_prepare_request() 915 memcpy(rctx->rembuf, rctx->buf, rctx->buf_cnt); in sahara_sha_prepare_request() 928 if (rctx->buf_cnt && req->nbytes) { in sahara_sha_prepare_request() 930 sg_set_buf(rctx->in_sg_chain, rctx->rembuf, rctx->buf_cnt); in sahara_sha_prepare_request() 934 } else if (rctx->buf_cnt) { in sahara_sha_prepare_request() 936 sg_init_one(rctx->in_sg, rctx->rembuf, rctx->buf_cnt); in sahara_sha_prepare_request() [all …]
|
/drivers/media/dvb-core/ |
D | dvb_vb2.c | 39 ctx->buf_cnt = *nbuffers; in _queue_setup() 344 ctx->buf_cnt = req->count; in dvb_vb2_reqbufs() 349 ctx->buf_cnt, ctx->buf_siz, ret); in dvb_vb2_reqbufs() 354 ctx->buf_cnt, ctx->buf_siz); in dvb_vb2_reqbufs()
|
/drivers/staging/rts5208/ |
D | rtsx_transport.c | 487 int buf_cnt, i; in rtsx_transfer_sglist_adma() local 521 buf_cnt = dma_map_sg(&rtsx->pci->dev, sg, num_sg, dma_dir); in rtsx_transfer_sglist_adma() 525 for (i = 0; i <= buf_cnt / (HOST_SG_TBL_BUF_LEN / 8); i++) { in rtsx_transfer_sglist_adma() 529 if (i == buf_cnt / (HOST_SG_TBL_BUF_LEN / 8)) in rtsx_transfer_sglist_adma() 530 sg_cnt = buf_cnt % (HOST_SG_TBL_BUF_LEN / 8); in rtsx_transfer_sglist_adma()
|
/drivers/net/ethernet/freescale/dpaa2/ |
D | dpaa2-ethtool.c | 235 u32 buf_cnt; in dpaa2_eth_get_ethtool_stats() local 309 err = dpaa2_io_query_bp_count(NULL, priv->bpid, &buf_cnt); in dpaa2_eth_get_ethtool_stats() 314 *(data + i++) = buf_cnt; in dpaa2_eth_get_ethtool_stats()
|
/drivers/net/ethernet/neterion/ |
D | s2io.c | 2718 int i, blk = 0, buf_cnt = 0; in free_rx_buffers() local 2734 dev->name, buf_cnt, i); in free_rx_buffers()
|