Searched refs:bd_count (Results 1 – 12 of 12) sorted by relevance
/drivers/net/ethernet/freescale/enetc/ |
D | enetc.c | 167 if (unlikely(i == tx_ring->bd_count)) { in enetc_map_tx_buffs() 204 if (unlikely(i == tx_ring->bd_count)) { in enetc_map_tx_buffs() 242 i = tx_ring->bd_count; in enetc_map_tx_buffs() 304 return pi >= ci ? pi - ci : tx_ring->bd_count - ci + pi; in enetc_bd_ready_count() 380 if (unlikely(i == tx_ring->bd_count)) { in enetc_clean_tx_ring() 462 if (unlikely(i == rx_ring->bd_count)) { in enetc_refill_rx_ring() 663 if (unlikely(i == rx_ring->bd_count)) { in enetc_clean_rx_ring() 676 if (unlikely(i == rx_ring->bd_count)) { in enetc_clean_rx_ring() 703 if (unlikely(i == rx_ring->bd_count)) { in enetc_clean_rx_ring() 752 r->bd_base = dma_alloc_coherent(r->dev, r->bd_count * bd_size, in enetc_dma_alloc_bdr() [all …]
|
D | enetc.h | 59 int bd_count; /* # of BDs */ member 79 if (unlikely(++*i == bdr->bd_count)) in enetc_bdr_idx_inc() 88 return bdr->bd_count + bdr->next_to_clean - bdr->next_to_use - 1; in enetc_bd_unused() 98 int bd_count; /* # of BDs */ member
|
D | enetc_cbdr.c | 23 i = (i + 1) % ring->bd_count; in enetc_clean_cbdr() 31 return (r->next_to_clean - r->next_to_use - 1 + r->bd_count) % in enetc_cbd_unused() 32 r->bd_count; in enetc_cbd_unused() 53 i = (i + 1) % ring->bd_count; in enetc_send_cmd()
|
/drivers/dma/bestcomm/ |
D | bestcomm.c | 50 bcom_task_alloc(int bd_count, int bd_size, int priv_size) in bcom_task_alloc() argument 89 if (bd_count) { in bcom_task_alloc() 90 tsk->cookie = kmalloc_array(bd_count, sizeof(void *), in bcom_task_alloc() 95 tsk->bd = bcom_sram_alloc(bd_count * bd_size, 4, &tsk->bd_pa); in bcom_task_alloc() 98 memset(tsk->bd, 0x00, bd_count * bd_size); in bcom_task_alloc() 100 tsk->num_bd = bd_count; in bcom_task_alloc()
|
/drivers/scsi/bnx2fc/ |
D | bnx2fc_io.c | 1653 int bd_count = 0; in bnx2fc_map_sg() local 1671 bd_count); in bnx2fc_map_sg() 1675 bd[bd_count].buf_addr_lo = addr & 0xffffffff; in bnx2fc_map_sg() 1676 bd[bd_count].buf_addr_hi = addr >> 32; in bnx2fc_map_sg() 1677 bd[bd_count].buf_len = (u16)sg_len; in bnx2fc_map_sg() 1678 bd[bd_count].flags = 0; in bnx2fc_map_sg() 1680 bd_count += sg_frags; in bnx2fc_map_sg() 1687 return bd_count; in bnx2fc_map_sg() 1694 int bd_count; in bnx2fc_build_bd_list_from_sg() local 1697 bd_count = bnx2fc_map_sg(io_req); in bnx2fc_build_bd_list_from_sg() [all …]
|
D | bnx2fc_hwi.c | 1465 int bd_count; in bnx2fc_init_seq_cleanup_task() local 1495 bd_count = orig_io_req->bd_tbl->bd_valid; in bnx2fc_init_seq_cleanup_task() 1498 for (i = 0; i < bd_count; i++) { in bnx2fc_init_seq_cleanup_task() 1511 bd_count; in bnx2fc_init_seq_cleanup_task() 1523 sgl->mul_sgl.sgl_size = bd_count; in bnx2fc_init_seq_cleanup_task() 1681 int bd_count; in bnx2fc_init_task() local 1694 bd_count = bd_tbl->bd_valid; in bnx2fc_init_task() 1697 if ((dev_type == TYPE_DISK) && (bd_count == 1)) { in bnx2fc_init_task() 1777 bd_count = bd_tbl->bd_valid; in bnx2fc_init_task() 1780 if (bd_count == 1) { in bnx2fc_init_task() [all …]
|
/drivers/scsi/bnx2i/ |
D | bnx2i_iscsi.c | 149 int bd_count = 0; in bnx2i_map_scsi_sg() local 162 bd[bd_count].buffer_addr_lo = addr & 0xffffffff; in bnx2i_map_scsi_sg() 163 bd[bd_count].buffer_addr_hi = addr >> 32; in bnx2i_map_scsi_sg() 164 bd[bd_count].buffer_length = sg_len; in bnx2i_map_scsi_sg() 165 bd[bd_count].flags = 0; in bnx2i_map_scsi_sg() 166 if (bd_count == 0) in bnx2i_map_scsi_sg() 167 bd[bd_count].flags = ISCSI_BD_FIRST_IN_BD_CHAIN; in bnx2i_map_scsi_sg() 170 bd_count++; in bnx2i_map_scsi_sg() 173 if (bd_count) in bnx2i_map_scsi_sg() 174 bd[bd_count - 1].flags |= ISCSI_BD_LAST_IN_BD_CHAIN; in bnx2i_map_scsi_sg() [all …]
|
/drivers/scsi/qedi/ |
D | qedi_fw.c | 1844 int bd_count = 0; in qedi_map_scsi_sg() local 1865 bd[bd_count].sge_addr.lo = (addr & 0xffffffff); in qedi_map_scsi_sg() 1866 bd[bd_count].sge_addr.hi = (addr >> 32); in qedi_map_scsi_sg() 1867 bd[bd_count].sge_len = (u16)sg_len; in qedi_map_scsi_sg() 1873 return ++bd_count; in qedi_map_scsi_sg() 1909 sg_frags = qedi_split_bd(cmd, addr, sg_len, bd_count); in qedi_map_scsi_sg() 1912 bd[bd_count].sge_addr.lo = addr & 0xffffffff; in qedi_map_scsi_sg() 1913 bd[bd_count].sge_addr.hi = addr >> 32; in qedi_map_scsi_sg() 1914 bd[bd_count].sge_len = sg_len; in qedi_map_scsi_sg() 1917 bd_count += sg_frags; in qedi_map_scsi_sg() [all …]
|
/drivers/scsi/qedf/ |
D | qedf_io.c | 488 int bd_count = 0; in qedf_map_sg() local 516 bd[bd_count].sge_addr.lo = cpu_to_le32(U64_LO(addr)); in qedf_map_sg() 517 bd[bd_count].sge_addr.hi = cpu_to_le32(U64_HI(addr)); in qedf_map_sg() 518 bd[bd_count].sge_len = cpu_to_le32(sg_len); in qedf_map_sg() 520 bd_count++; in qedf_map_sg() 533 return bd_count; in qedf_map_sg() 540 int bd_count; in qedf_build_bd_list_from_sg() local 543 bd_count = qedf_map_sg(io_req); in qedf_build_bd_list_from_sg() 544 if (bd_count == 0) in qedf_build_bd_list_from_sg() 547 bd_count = 0; in qedf_build_bd_list_from_sg() [all …]
|
/drivers/net/ethernet/toshiba/ |
D | tc35815.c | 1072 int i, bd_count = (le32_to_cpu(fd->fd.FDCtl) & FD_BDCnt_MASK) >> FD_BDCnt_SHIFT; in dump_rxfd() local 1073 if (bd_count > 8) in dump_rxfd() 1074 bd_count = 8; in dump_rxfd() 1083 for (i = 0; i < bd_count; i++) in dump_rxfd() 1088 return bd_count; in dump_rxfd() 1123 int bd_count = dump_rxfd(&lp->rfd_base[i]); in panic_queues() local 1124 i += (bd_count + 1) / 2; /* skip BDs */ in panic_queues() 1467 int bd_count = (fdctl & FD_BDCnt_MASK) >> FD_BDCnt_SHIFT; in tc35815_rx() local 1484 BUG_ON(bd_count > 1); in tc35815_rx() 1538 if (bd_count > 0) { in tc35815_rx() [all …]
|
/drivers/block/zram/ |
D | zram_drv.h | 86 atomic64_t bd_count; /* no. of pages in backing device */ member
|
D | zram_drv.c | 564 atomic64_inc(&zram->stats.bd_count); in alloc_block_bdev() 574 atomic64_dec(&zram->stats.bd_count); in free_block_bdev() 1094 FOUR_K((u64)atomic64_read(&zram->stats.bd_count)), in bd_stat_show()
|