Home
last modified time | relevance | path

Searched refs:buff_size (Results 1 – 25 of 46) sorted by relevance

12

/drivers/firmware/efi/libstub/
Dmem.c8 static inline bool mmap_has_headroom(unsigned long buff_size, in mmap_has_headroom() argument
12 unsigned long slack = buff_size - map_size; in mmap_has_headroom()
35 *map->buff_size = *map->map_size; in efi_get_memory_map()
47 !mmap_has_headroom(*map->buff_size, *map->map_size, in efi_get_memory_map()
58 *map->buff_size = *map->map_size; in efi_get_memory_map()
Drandomalloc.c59 unsigned long buff_size; in efi_random_alloc() local
70 map.buff_size = &buff_size; in efi_random_alloc()
Drelocate.c26 unsigned long map_size, desc_size, buff_size; in efi_low_alloc_above() local
38 boot_map.buff_size = &buff_size; in efi_low_alloc_above()
Darm64-stub.c45 unsigned long map_size, desc_size, buff_size; in check_image_region() local
57 map.buff_size = &buff_size; in check_image_region()
Dfdt.c247 unsigned long map_size, desc_size, buff_size; in allocate_new_fdt_and_exit_boot() local
261 map.buff_size = &buff_size; in allocate_new_fdt_and_exit_boot()
Dx86-stub.c634 unsigned long map_sz, key, desc_size, buff_size; in exit_boot() local
648 map.buff_size = &buff_size; in exit_boot()
/drivers/net/ethernet/ibm/
Dibmveth.c147 u32 buff_size, u32 pool_active) in ibmveth_init_buffer_pool() argument
151 pool->buff_size = buff_size; in ibmveth_init_buffer_pool()
222 skb = netdev_alloc_skb(adapter->netdev, pool->buff_size); in ibmveth_replenish_buffer_pool()
241 pool->buff_size, DMA_FROM_DEVICE); in ibmveth_replenish_buffer_pool()
253 desc.fields.flags_len = IBMVETH_BUF_VALID | pool->buff_size; in ibmveth_replenish_buffer_pool()
257 unsigned int len = min(pool->buff_size, in ibmveth_replenish_buffer_pool()
286 pool->dma_addr[index], pool->buff_size, in ibmveth_replenish_buffer_pool()
340 pool->buff_size, in ibmveth_free_buffer_pool()
379 adapter->rx_buff_pool[pool].buff_size, in ibmveth_remove_buffer_from_pool()
428 adapter->rx_buff_pool[pool].buff_size; in ibmveth_rxq_recycle_buffer()
[all …]
Dibmvnic.c333 skb = alloc_skb(pool->buff_size, GFP_ATOMIC); in replenish_rx_pool()
346 offset = index * pool->buff_size; in replenish_rx_pool()
348 memset(dst, 0, pool->buff_size); in replenish_rx_pool()
356 pool->rx_buff[index].size = pool->buff_size; in replenish_rx_pool()
373 sub_crq.rx_add.len = cpu_to_be32(pool->buff_size << shift); in replenish_rx_pool()
481 u64 buff_size; in reset_rx_pools() local
488 buff_size = adapter->cur_rx_buf_sz; in reset_rx_pools()
495 if (rx_pool->buff_size != buff_size) { in reset_rx_pools()
497 rx_pool->buff_size = buff_size; in reset_rx_pools()
501 rx_pool->buff_size); in reset_rx_pools()
[all …]
Dibmveth.h112 u32 buff_size; member
/drivers/net/ethernet/aquantia/atlantic/hw_atl/
Dhw_atl_a0.c128 u32 buff_size = 0U; in hw_atl_a0_hw_qos_set() local
148 buff_size = HW_ATL_A0_TXBUF_MAX; in hw_atl_a0_hw_qos_set()
150 hw_atl_tpb_tx_pkt_buff_size_per_tc_set(self, buff_size, tc); in hw_atl_a0_hw_qos_set()
152 (buff_size * in hw_atl_a0_hw_qos_set()
156 (buff_size * in hw_atl_a0_hw_qos_set()
163 buff_size = HW_ATL_A0_RXBUF_MAX; in hw_atl_a0_hw_qos_set()
165 hw_atl_rpb_rx_pkt_buff_size_per_tc_set(self, buff_size, tc); in hw_atl_a0_hw_qos_set()
167 (buff_size * in hw_atl_a0_hw_qos_set()
171 (buff_size * in hw_atl_a0_hw_qos_set()
/drivers/net/ethernet/intel/iavf/
Diavf_adminq.c646 u16 buff_size, in iavf_asq_send_command() argument
698 if (buff_size > hw->aq.asq_buf_size) { in iavf_asq_send_command()
702 buff_size); in iavf_asq_send_command()
740 memcpy(dma_buff->va, buff, buff_size); in iavf_asq_send_command()
741 desc_on_ring->datalen = cpu_to_le16(buff_size); in iavf_asq_send_command()
755 buff, buff_size); in iavf_asq_send_command()
783 memcpy(buff, dma_buff->va, buff_size); in iavf_asq_send_command()
806 iavf_debug_aq(hw, IAVF_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size); in iavf_asq_send_command()
Diavf_prototype.h28 u16 buff_size,
/drivers/net/ethernet/intel/i40e/
Di40e_prototype.h28 u16 buff_size,
189 void *buff, u16 buff_size, u16 *data_size,
200 u8 mib_type, void *buff, u16 buff_size,
219 void *buff, u16 buff_size,
400 u8 table_id, u32 start_index, u16 buff_size,
449 u16 buff_size, u32 track_id,
454 u16 buff_size, u8 flags,
Di40e_adminq.c785 u16 buff_size, in i40e_asq_send_command() argument
837 if (buff_size > hw->aq.asq_buf_size) { in i40e_asq_send_command()
841 buff_size); in i40e_asq_send_command()
879 memcpy(dma_buff->va, buff, buff_size); in i40e_asq_send_command()
880 desc_on_ring->datalen = cpu_to_le16(buff_size); in i40e_asq_send_command()
894 buff, buff_size); in i40e_asq_send_command()
922 memcpy(buff, dma_buff->va, buff_size); in i40e_asq_send_command()
945 i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size); in i40e_asq_send_command()
Di40e_common.c3483 void *buff, u16 buff_size, u16 *data_size, in i40e_aq_discover_capabilities() argument
3502 if (buff_size > I40E_AQ_LARGE_BUF) in i40e_aq_discover_capabilities()
3505 status = i40e_asq_send_command(hw, &desc, buff, buff_size, cmd_details); in i40e_aq_discover_capabilities()
3625 u8 mib_type, void *buff, u16 buff_size, in i40e_aq_get_lldp_mib() argument
3636 if (buff_size == 0 || !buff) in i40e_aq_get_lldp_mib()
3647 desc.datalen = cpu_to_le16(buff_size); in i40e_aq_get_lldp_mib()
3650 if (buff_size > I40E_AQ_LARGE_BUF) in i40e_aq_get_lldp_mib()
3653 status = i40e_asq_send_command(hw, &desc, buff, buff_size, cmd_details); in i40e_aq_get_lldp_mib()
3840 void *buff, u16 buff_size, in i40e_aq_get_cee_dcb_config() argument
3846 if (buff_size == 0 || !buff) in i40e_aq_get_cee_dcb_config()
[all …]
Di40e_nvm.c1379 u32 buff_size = 0; in i40e_nvmupd_exec_aq() local
1406 buff_size = max_t(u32, aq_data_len, le16_to_cpu(aq_desc->datalen)); in i40e_nvmupd_exec_aq()
1407 if (buff_size) { in i40e_nvmupd_exec_aq()
1428 buff_size, &cmd_details); in i40e_nvmupd_exec_aq()
/drivers/usb/renesas_usbhs/
Dpipe.c474 u16 buff_size; in usbhsp_setup_pipebuff() local
489 buff_size = pipe_config->bufsize; in usbhsp_setup_pipebuff()
493 bufnmb_cnt = (buff_size / 64) - 1; in usbhsp_setup_pipebuff()
496 pipe_num, buff_size, bufnmb); in usbhsp_setup_pipebuff()
/drivers/net/ethernet/qlogic/qed/
Dqed_iwarp.c1582 (u16)buf->buff_size, buf, 1); in qed_iwarp_ll2_post_rx()
1587 dma_free_coherent(&p_hwfn->cdev->pdev->dev, buf->buff_size, in qed_iwarp_ll2_post_rx()
1903 if ((fpdu->mpa_frag_len + tcp_payload_size) > (u16)buf->buff_size) { in qed_iwarp_cp_pkt()
1906 buf->buff_size, fpdu->mpa_frag_len, in qed_iwarp_cp_pkt()
1942 buf->buff_size, fpdu->mpa_frag_len, tcp_payload_size, in qed_iwarp_cp_pkt()
2474 dma_free_coherent(&p_hwfn->cdev->pdev->dev, buffer->buff_size, in qed_iwarp_ll2_rel_rx_pkt()
2517 buffer->piggy_buf->buff_size, in qed_iwarp_ll2_rel_tx_pkt()
2524 dma_free_coherent(&p_hwfn->cdev->pdev->dev, buffer->buff_size, in qed_iwarp_ll2_rel_tx_pkt()
2598 int num_rx_bufs, int buff_size, u8 ll2_handle) in qed_iwarp_ll2_alloc_buffers() argument
2612 buff_size, in qed_iwarp_ll2_alloc_buffers()
[all …]
Dqed_iwarp.h35 u32 buff_size; member
/drivers/infiniband/hw/hns/
Dhns_roce_qp.c501 hr_qp->buff_size = 0; in set_wqe_buf_attr()
511 hr_qp->buff_size += buf_size; in set_wqe_buf_attr()
515 hr_qp->sge.offset = hr_qp->buff_size; in set_wqe_buf_attr()
522 hr_qp->buff_size += buf_size; in set_wqe_buf_attr()
526 hr_qp->rq.offset = hr_qp->buff_size; in set_wqe_buf_attr()
533 hr_qp->buff_size += buf_size; in set_wqe_buf_attr()
536 if (hr_qp->buff_size < 1) in set_wqe_buf_attr()
/drivers/ntb/
Dntb_transport.c209 size_t buff_size; member
805 mw->buff_size = 0; in ntb_free_mw()
834 if (mw->alloc_size > mw->buff_size) { in ntb_alloc_mw_buffer()
860 size_t xlat_size, buff_size; in ntb_set_mw() local
874 buff_size = round_up(size, xlat_align); in ntb_set_mw()
880 if (mw->buff_size) in ntb_set_mw()
885 mw->buff_size = buff_size; in ntb_set_mw()
886 mw->alloc_size = buff_size; in ntb_set_mw()
896 mw->buff_size = 0; in ntb_set_mw()
1332 mw->buff_size = 0; in ntb_transport_probe()
/drivers/net/ethernet/aquantia/atlantic/
Daq_nic.c651 unsigned int buff_size = 0U; in aq_nic_map_skb() local
659 buff_size = AQ_CFG_TX_FRAME_MAX; in aq_nic_map_skb()
661 buff_size = frag_len; in aq_nic_map_skb()
666 buff_size, in aq_nic_map_skb()
677 dx_buff->len = buff_size; in aq_nic_map_skb()
682 frag_len -= buff_size; in aq_nic_map_skb()
683 buff_offset += buff_size; in aq_nic_map_skb()
/drivers/dma/
Dmv_xor_v2.c129 u32 buff_size; member
421 hw_descriptor->buff_size = len; in mv_xor_v2_prep_dma_memcpy()
480 hw_descriptor->buff_size = len; in mv_xor_v2_prep_dma_xor()
/drivers/gpu/drm/amd/display/dc/dsc/
Ddc_dsc.c768 int buff_size; in dc_dsc_parse_dsc_dpcd() local
773 buff_size = dpcd_dsc_basic_data[DP_DSC_RC_BUF_SIZE - DP_DSC_SUPPORT] + 1; in dc_dsc_parse_dsc_dpcd()
774 dsc_sink_caps->rc_buffer_size = buff_size * buff_block_size; in dc_dsc_parse_dsc_dpcd()
/drivers/crypto/ccree/
Dcc_buffer_mgr.c117 u32 buff_size, u32 *curr_nents, in cc_render_buff_to_mlli() argument
124 new_nents = (*curr_nents + buff_size / CC_MAX_MLLI_ENTRY_SIZE + 1); in cc_render_buff_to_mlli()
132 while (buff_size > CC_MAX_MLLI_ENTRY_SIZE) { in cc_render_buff_to_mlli()
139 buff_size -= CC_MAX_MLLI_ENTRY_SIZE; in cc_render_buff_to_mlli()
145 cc_lli_set_size(mlli_entry_p, buff_size); in cc_render_buff_to_mlli()

12