/drivers/firmware/efi/libstub/ |
D | mem.c | 8 static inline bool mmap_has_headroom(unsigned long buff_size, in mmap_has_headroom() argument 12 unsigned long slack = buff_size - map_size; in mmap_has_headroom() 35 *map->buff_size = *map->map_size; in efi_get_memory_map() 47 !mmap_has_headroom(*map->buff_size, *map->map_size, in efi_get_memory_map() 58 *map->buff_size = *map->map_size; in efi_get_memory_map()
|
D | randomalloc.c | 59 unsigned long buff_size; in efi_random_alloc() local 70 map.buff_size = &buff_size; in efi_random_alloc()
|
D | relocate.c | 26 unsigned long map_size, desc_size, buff_size; in efi_low_alloc_above() local 38 boot_map.buff_size = &buff_size; in efi_low_alloc_above()
|
D | arm64-stub.c | 45 unsigned long map_size, desc_size, buff_size; in check_image_region() local 57 map.buff_size = &buff_size; in check_image_region()
|
D | fdt.c | 247 unsigned long map_size, desc_size, buff_size; in allocate_new_fdt_and_exit_boot() local 261 map.buff_size = &buff_size; in allocate_new_fdt_and_exit_boot()
|
D | x86-stub.c | 634 unsigned long map_sz, key, desc_size, buff_size; in exit_boot() local 648 map.buff_size = &buff_size; in exit_boot()
|
/drivers/net/ethernet/ibm/ |
D | ibmveth.c | 147 u32 buff_size, u32 pool_active) in ibmveth_init_buffer_pool() argument 151 pool->buff_size = buff_size; in ibmveth_init_buffer_pool() 222 skb = netdev_alloc_skb(adapter->netdev, pool->buff_size); in ibmveth_replenish_buffer_pool() 241 pool->buff_size, DMA_FROM_DEVICE); in ibmveth_replenish_buffer_pool() 253 desc.fields.flags_len = IBMVETH_BUF_VALID | pool->buff_size; in ibmveth_replenish_buffer_pool() 257 unsigned int len = min(pool->buff_size, in ibmveth_replenish_buffer_pool() 286 pool->dma_addr[index], pool->buff_size, in ibmveth_replenish_buffer_pool() 340 pool->buff_size, in ibmveth_free_buffer_pool() 379 adapter->rx_buff_pool[pool].buff_size, in ibmveth_remove_buffer_from_pool() 428 adapter->rx_buff_pool[pool].buff_size; in ibmveth_rxq_recycle_buffer() [all …]
|
D | ibmvnic.c | 333 skb = alloc_skb(pool->buff_size, GFP_ATOMIC); in replenish_rx_pool() 346 offset = index * pool->buff_size; in replenish_rx_pool() 348 memset(dst, 0, pool->buff_size); in replenish_rx_pool() 356 pool->rx_buff[index].size = pool->buff_size; in replenish_rx_pool() 373 sub_crq.rx_add.len = cpu_to_be32(pool->buff_size << shift); in replenish_rx_pool() 481 u64 buff_size; in reset_rx_pools() local 488 buff_size = adapter->cur_rx_buf_sz; in reset_rx_pools() 495 if (rx_pool->buff_size != buff_size) { in reset_rx_pools() 497 rx_pool->buff_size = buff_size; in reset_rx_pools() 501 rx_pool->buff_size); in reset_rx_pools() [all …]
|
D | ibmveth.h | 112 u32 buff_size; member
|
/drivers/net/ethernet/aquantia/atlantic/hw_atl/ |
D | hw_atl_a0.c | 128 u32 buff_size = 0U; in hw_atl_a0_hw_qos_set() local 148 buff_size = HW_ATL_A0_TXBUF_MAX; in hw_atl_a0_hw_qos_set() 150 hw_atl_tpb_tx_pkt_buff_size_per_tc_set(self, buff_size, tc); in hw_atl_a0_hw_qos_set() 152 (buff_size * in hw_atl_a0_hw_qos_set() 156 (buff_size * in hw_atl_a0_hw_qos_set() 163 buff_size = HW_ATL_A0_RXBUF_MAX; in hw_atl_a0_hw_qos_set() 165 hw_atl_rpb_rx_pkt_buff_size_per_tc_set(self, buff_size, tc); in hw_atl_a0_hw_qos_set() 167 (buff_size * in hw_atl_a0_hw_qos_set() 171 (buff_size * in hw_atl_a0_hw_qos_set()
|
/drivers/net/ethernet/intel/iavf/ |
D | iavf_adminq.c | 646 u16 buff_size, in iavf_asq_send_command() argument 698 if (buff_size > hw->aq.asq_buf_size) { in iavf_asq_send_command() 702 buff_size); in iavf_asq_send_command() 740 memcpy(dma_buff->va, buff, buff_size); in iavf_asq_send_command() 741 desc_on_ring->datalen = cpu_to_le16(buff_size); in iavf_asq_send_command() 755 buff, buff_size); in iavf_asq_send_command() 783 memcpy(buff, dma_buff->va, buff_size); in iavf_asq_send_command() 806 iavf_debug_aq(hw, IAVF_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size); in iavf_asq_send_command()
|
D | iavf_prototype.h | 28 u16 buff_size,
|
/drivers/net/ethernet/intel/i40e/ |
D | i40e_prototype.h | 28 u16 buff_size, 189 void *buff, u16 buff_size, u16 *data_size, 200 u8 mib_type, void *buff, u16 buff_size, 219 void *buff, u16 buff_size, 400 u8 table_id, u32 start_index, u16 buff_size, 449 u16 buff_size, u32 track_id, 454 u16 buff_size, u8 flags,
|
D | i40e_adminq.c | 785 u16 buff_size, in i40e_asq_send_command() argument 837 if (buff_size > hw->aq.asq_buf_size) { in i40e_asq_send_command() 841 buff_size); in i40e_asq_send_command() 879 memcpy(dma_buff->va, buff, buff_size); in i40e_asq_send_command() 880 desc_on_ring->datalen = cpu_to_le16(buff_size); in i40e_asq_send_command() 894 buff, buff_size); in i40e_asq_send_command() 922 memcpy(buff, dma_buff->va, buff_size); in i40e_asq_send_command() 945 i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size); in i40e_asq_send_command()
|
D | i40e_common.c | 3483 void *buff, u16 buff_size, u16 *data_size, in i40e_aq_discover_capabilities() argument 3502 if (buff_size > I40E_AQ_LARGE_BUF) in i40e_aq_discover_capabilities() 3505 status = i40e_asq_send_command(hw, &desc, buff, buff_size, cmd_details); in i40e_aq_discover_capabilities() 3625 u8 mib_type, void *buff, u16 buff_size, in i40e_aq_get_lldp_mib() argument 3636 if (buff_size == 0 || !buff) in i40e_aq_get_lldp_mib() 3647 desc.datalen = cpu_to_le16(buff_size); in i40e_aq_get_lldp_mib() 3650 if (buff_size > I40E_AQ_LARGE_BUF) in i40e_aq_get_lldp_mib() 3653 status = i40e_asq_send_command(hw, &desc, buff, buff_size, cmd_details); in i40e_aq_get_lldp_mib() 3840 void *buff, u16 buff_size, in i40e_aq_get_cee_dcb_config() argument 3846 if (buff_size == 0 || !buff) in i40e_aq_get_cee_dcb_config() [all …]
|
D | i40e_nvm.c | 1379 u32 buff_size = 0; in i40e_nvmupd_exec_aq() local 1406 buff_size = max_t(u32, aq_data_len, le16_to_cpu(aq_desc->datalen)); in i40e_nvmupd_exec_aq() 1407 if (buff_size) { in i40e_nvmupd_exec_aq() 1428 buff_size, &cmd_details); in i40e_nvmupd_exec_aq()
|
/drivers/usb/renesas_usbhs/ |
D | pipe.c | 474 u16 buff_size; in usbhsp_setup_pipebuff() local 489 buff_size = pipe_config->bufsize; in usbhsp_setup_pipebuff() 493 bufnmb_cnt = (buff_size / 64) - 1; in usbhsp_setup_pipebuff() 496 pipe_num, buff_size, bufnmb); in usbhsp_setup_pipebuff()
|
/drivers/net/ethernet/qlogic/qed/ |
D | qed_iwarp.c | 1582 (u16)buf->buff_size, buf, 1); in qed_iwarp_ll2_post_rx() 1587 dma_free_coherent(&p_hwfn->cdev->pdev->dev, buf->buff_size, in qed_iwarp_ll2_post_rx() 1903 if ((fpdu->mpa_frag_len + tcp_payload_size) > (u16)buf->buff_size) { in qed_iwarp_cp_pkt() 1906 buf->buff_size, fpdu->mpa_frag_len, in qed_iwarp_cp_pkt() 1942 buf->buff_size, fpdu->mpa_frag_len, tcp_payload_size, in qed_iwarp_cp_pkt() 2474 dma_free_coherent(&p_hwfn->cdev->pdev->dev, buffer->buff_size, in qed_iwarp_ll2_rel_rx_pkt() 2517 buffer->piggy_buf->buff_size, in qed_iwarp_ll2_rel_tx_pkt() 2524 dma_free_coherent(&p_hwfn->cdev->pdev->dev, buffer->buff_size, in qed_iwarp_ll2_rel_tx_pkt() 2598 int num_rx_bufs, int buff_size, u8 ll2_handle) in qed_iwarp_ll2_alloc_buffers() argument 2612 buff_size, in qed_iwarp_ll2_alloc_buffers() [all …]
|
D | qed_iwarp.h | 35 u32 buff_size; member
|
/drivers/infiniband/hw/hns/ |
D | hns_roce_qp.c | 501 hr_qp->buff_size = 0; in set_wqe_buf_attr() 511 hr_qp->buff_size += buf_size; in set_wqe_buf_attr() 515 hr_qp->sge.offset = hr_qp->buff_size; in set_wqe_buf_attr() 522 hr_qp->buff_size += buf_size; in set_wqe_buf_attr() 526 hr_qp->rq.offset = hr_qp->buff_size; in set_wqe_buf_attr() 533 hr_qp->buff_size += buf_size; in set_wqe_buf_attr() 536 if (hr_qp->buff_size < 1) in set_wqe_buf_attr()
|
/drivers/ntb/ |
D | ntb_transport.c | 209 size_t buff_size; member 805 mw->buff_size = 0; in ntb_free_mw() 834 if (mw->alloc_size > mw->buff_size) { in ntb_alloc_mw_buffer() 860 size_t xlat_size, buff_size; in ntb_set_mw() local 874 buff_size = round_up(size, xlat_align); in ntb_set_mw() 880 if (mw->buff_size) in ntb_set_mw() 885 mw->buff_size = buff_size; in ntb_set_mw() 886 mw->alloc_size = buff_size; in ntb_set_mw() 896 mw->buff_size = 0; in ntb_set_mw() 1332 mw->buff_size = 0; in ntb_transport_probe()
|
/drivers/net/ethernet/aquantia/atlantic/ |
D | aq_nic.c | 651 unsigned int buff_size = 0U; in aq_nic_map_skb() local 659 buff_size = AQ_CFG_TX_FRAME_MAX; in aq_nic_map_skb() 661 buff_size = frag_len; in aq_nic_map_skb() 666 buff_size, in aq_nic_map_skb() 677 dx_buff->len = buff_size; in aq_nic_map_skb() 682 frag_len -= buff_size; in aq_nic_map_skb() 683 buff_offset += buff_size; in aq_nic_map_skb()
|
/drivers/dma/ |
D | mv_xor_v2.c | 129 u32 buff_size; member 421 hw_descriptor->buff_size = len; in mv_xor_v2_prep_dma_memcpy() 480 hw_descriptor->buff_size = len; in mv_xor_v2_prep_dma_xor()
|
/drivers/gpu/drm/amd/display/dc/dsc/ |
D | dc_dsc.c | 768 int buff_size; in dc_dsc_parse_dsc_dpcd() local 773 buff_size = dpcd_dsc_basic_data[DP_DSC_RC_BUF_SIZE - DP_DSC_SUPPORT] + 1; in dc_dsc_parse_dsc_dpcd() 774 dsc_sink_caps->rc_buffer_size = buff_size * buff_block_size; in dc_dsc_parse_dsc_dpcd()
|
/drivers/crypto/ccree/ |
D | cc_buffer_mgr.c | 117 u32 buff_size, u32 *curr_nents, in cc_render_buff_to_mlli() argument 124 new_nents = (*curr_nents + buff_size / CC_MAX_MLLI_ENTRY_SIZE + 1); in cc_render_buff_to_mlli() 132 while (buff_size > CC_MAX_MLLI_ENTRY_SIZE) { in cc_render_buff_to_mlli() 139 buff_size -= CC_MAX_MLLI_ENTRY_SIZE; in cc_render_buff_to_mlli() 145 cc_lli_set_size(mlli_entry_p, buff_size); in cc_render_buff_to_mlli()
|