/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ice/ |
D | ice_fltr.c | 57 return ice_add_mac(&vsi->back->hw, list); in ice_fltr_add_mac_list() 68 return ice_remove_mac(&vsi->back->hw, list); in ice_fltr_remove_mac_list() 79 return ice_add_vlan(&vsi->back->hw, list); in ice_fltr_add_vlan_list() 90 return ice_remove_vlan(&vsi->back->hw, list); in ice_fltr_remove_vlan_list() 101 return ice_add_eth_mac(&vsi->back->hw, list); in ice_fltr_add_eth_list() 112 return ice_remove_eth_mac(&vsi->back->hw, list); in ice_fltr_remove_eth_list() 121 ice_remove_vsi_fltr(&vsi->back->hw, vsi->idx); in ice_fltr_remove_all() 145 return ice_fltr_add_entry_to_list(ice_pf_to_dev(vsi->back), &info, in ice_fltr_add_mac_to_list() 169 return ice_fltr_add_entry_to_list(ice_pf_to_dev(vsi->back), &info, in ice_fltr_add_vlan_to_list() 199 return ice_fltr_add_entry_to_list(ice_pf_to_dev(vsi->back), &info, in ice_fltr_add_eth_to_list() [all …]
|
D | ice_lib.c | 50 ice_flush(&vsi->back->hw); in ice_vsi_ctrl_all_rx_rings() 70 struct ice_pf *pf = vsi->back; in ice_vsi_alloc_arrays() 146 dev_dbg(ice_pf_to_dev(vsi->back), "Not setting number of Tx/Rx descriptors for VSI type %d\n", in ice_vsi_set_num_desc() 161 struct ice_pf *pf = vsi->back; in ice_vsi_set_num_qs() 263 struct ice_pf *pf = vsi->back; in ice_vsi_delete() 291 struct ice_pf *pf = vsi->back; in ice_vsi_free_arrays() 336 if (!vsi->back) in ice_vsi_clear() 339 pf = vsi->back; in ice_vsi_clear() 428 vsi->back = pf; in ice_vsi_alloc() 498 struct ice_pf *pf = vsi->back; in ice_alloc_fd_res() [all …]
|
D | ice_base.c | 103 struct ice_pf *pf = vsi->back; in ice_vsi_alloc_q_vector() 145 struct ice_pf *pf = vsi->back; in ice_free_q_vector() 229 struct ice_hw *hw = &vsi->back->hw; in ice_setup_tx_ctx() 285 struct device *dev = ice_pf_to_dev(ring->vsi->back); in ice_setup_rx_ctx() 295 hw = &vsi->back->hw; in ice_setup_rx_ctx() 412 if (!vsi->netdev || test_bit(ICE_FLAG_LEGACY_RX, vsi->back->flags)) in ice_setup_rx_ctx() 479 struct ice_pf *pf = vsi->back; in ice_vsi_ctrl_one_rx_ring() 517 struct ice_pf *pf = vsi->back; in ice_vsi_wait_one_rx_ring() 531 struct device *dev = ice_pf_to_dev(vsi->back); in ice_vsi_alloc_q_vectors() 642 struct ice_pf *pf = vsi->back; in ice_vsi_cfg_txq() [all …]
|
D | ice_arfs.c | 27 struct ice_pf *pf = hw->back; in ice_is_arfs_using_perfect_flow() 90 …dev_err(ice_pf_to_dev(vsi->back), "aRFS: Failed to update filter counters, invalid filter type %d\… in ice_arfs_update_active_fltr_cntrs() 111 dev = ice_pf_to_dev(vsi->back); in ice_arfs_del_flow_rules() 116 result = ice_fdir_write_fltr(vsi->back, &e->fltr_info, false, in ice_arfs_del_flow_rules() 148 dev = ice_pf_to_dev(vsi->back); in ice_arfs_add_flow_rules() 153 result = ice_fdir_write_fltr(vsi->back, in ice_arfs_add_flow_rules() 222 dev = ice_pf_to_dev(vsi->back); in ice_arfs_update_flow_rules() 307 arfs_entry = devm_kzalloc(ice_pf_to_dev(vsi->back), in ice_arfs_build_entry() 413 pf = vsi->back; in ice_rx_flow_steer() 546 if (!vsi || vsi->type != ICE_VSI_PF || !vsi->back || in ice_clear_arfs() [all …]
|
/kernel/linux/linux-5.10/drivers/net/ethernet/netronome/nfp/crypto/ |
D | tls.c | 189 return &req->back; in nfp_net_tls_set_ipv4() 209 return &req->back; in nfp_net_tls_set_ipv6() 214 struct nfp_crypto_req_add_back *back, struct sock *sk, in nfp_net_tls_set_l4() argument 222 back->src_port = 0; in nfp_net_tls_set_l4() 223 back->dst_port = 0; in nfp_net_tls_set_l4() 225 back->src_port = inet->inet_dport; in nfp_net_tls_set_l4() 226 back->dst_port = inet->inet_sport; in nfp_net_tls_set_l4() 273 struct nfp_crypto_req_add_back *back; in nfp_net_tls_add() local 328 back = nfp_net_tls_set_ipv6(nn, req, sk, direction); in nfp_net_tls_add() 330 back = nfp_net_tls_set_ipv4(nn, req, sk, direction); in nfp_net_tls_add() [all …]
|
D | fw.h | 61 struct nfp_crypto_req_add_back back; member 68 struct nfp_crypto_req_add_back back; member
|
/kernel/linux/linux-5.10/fs/erofs/ |
D | namei.c | 51 int head, back; in find_target_dirent() local 57 back = ndirents - 1; in find_target_dirent() 60 while (head <= back) { in find_target_dirent() 61 const int mid = head + (back - head) / 2; in find_target_dirent() 82 back = mid - 1; in find_target_dirent() 95 int head, back; in find_target_block_classic() local 101 back = erofs_inode_datablocks(dir) - 1; in find_target_block_classic() 103 while (head <= back) { in find_target_block_classic() 104 const int mid = head + (back - head) / 2; in find_target_block_classic() 155 back = mid - 1; in find_target_block_classic()
|
/kernel/linux/linux-5.10/Documentation/scsi/ |
D | sd-parameters.rst | 16 write back 1 0 on on 17 write back, no read (daft) 1 1 on off 20 To set cache type to "write back" and save this setting to the drive:: 22 # echo "write back" > cache_type 27 # echo "temporary write back" > cache_type
|
/kernel/linux/linux-5.10/lib/lz4/ |
D | lz4hc_compress.c | 208 int back = 0; in LZ4HC_InsertAndGetWiderMatch() local 210 while ((ip + back > iLowLimit) in LZ4HC_InsertAndGetWiderMatch() 211 && (matchPtr + back > lowPrefixPtr) in LZ4HC_InsertAndGetWiderMatch() 212 && (ip[back - 1] == matchPtr[back - 1])) in LZ4HC_InsertAndGetWiderMatch() 213 back--; in LZ4HC_InsertAndGetWiderMatch() 215 mlt -= back; in LZ4HC_InsertAndGetWiderMatch() 219 *matchpos = matchPtr + back; in LZ4HC_InsertAndGetWiderMatch() 220 *startpos = ip + back; in LZ4HC_InsertAndGetWiderMatch() 229 int back = 0; in LZ4HC_InsertAndGetWiderMatch() local 241 while ((ip + back > iLowLimit) in LZ4HC_InsertAndGetWiderMatch() [all …]
|
/kernel/linux/linux-5.10/Documentation/driver-api/md/ |
D | raid5-cache.rst | 8 since 4.4) or write-back mode (supported since 4.10). mdadm (supported since 11 in write-through mode. A user can switch it to write-back mode by:: 13 echo "write-back" > /sys/block/md0/md/journal_mode 15 And switch it back to write-through mode by:: 30 tries to resync the array to bring it back to normal state. But before the 47 write-back mode 50 write-back mode fixes the 'write hole' issue too, since all write data is 51 cached on cache disk. But the main goal of 'write-back' cache is to speed up 56 overhead too. Write-back cache will aggregate the data and flush the data to 61 In write-back mode, MD reports IO completion to upper layer (usually [all …]
|
/kernel/linux/linux-5.10/scripts/coccinelle/iterators/ |
D | list_entry_update.cocci | 5 /// that there is a path from the reassignment back to the top of the loop. 39 @back depends on (org || report) && !context exists@ 49 @script:python depends on back && org@ 57 @script:python depends on back && report@
|
/kernel/linux/linux-5.10/drivers/md/bcache/ |
D | util.h | 118 size_t front, back, size, mask; \ 124 c = (fifo)->data[iter], iter != (fifo)->back; \ 136 (fifo)->front = (fifo)->back = 0; \ 162 #define fifo_used(fifo) (((fifo)->back - (fifo)->front) & (fifo)->mask) 170 ((fifo)->data[((fifo)->back - 1) & (fifo)->mask]) 178 (fifo)->data[(fifo)->back++] = (i); \ 179 (fifo)->back &= (fifo)->mask; \ 209 --(fifo)->back; \ 210 (fifo)->back &= (fifo)->mask; \ 211 (i) = (fifo)->data[(fifo)->back] \ [all …]
|
/kernel/linux/linux-5.10/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
D | hclge_mbx.c | 35 struct hclge_dev *hdev = vport->back; in hclge_gen_resp_to_vf() 90 struct hclge_dev *hdev = vport->back; in hclge_send_mbx_msg() 124 struct hclge_dev *hdev = vport->back; in hclge_inform_reset_assert_to_vf() 176 struct hclge_dev *hdev = vport->back; in hclge_get_ring_chain_from_mbx() 286 struct hclge_dev *hdev = vport->back; in hclge_set_vf_uc_mac_addr() 328 struct hclge_dev *hdev = vport->back; in hclge_set_vf_mc_mac_addr() 440 struct hclge_dev *hdev = vport->back; in hclge_get_vf_queue_info() 466 struct hclge_dev *hdev = vport->back; in hclge_get_vf_queue_depth() 483 struct hclge_dev *hdev = vport->back; in hclge_get_vf_media_type() 498 struct hclge_dev *hdev = vport->back; in hclge_get_link_info() [all …]
|
D | hclge_dcb.c | 73 struct hclge_dev *hdev = vport->back; in hclge_ieee_getets() 272 struct hclge_dev *hdev = vport->back; in hclge_ieee_setets() 332 struct hclge_dev *hdev = vport->back; in hclge_ieee_getpfc() 359 struct hclge_dev *hdev = vport->back; in hclge_ieee_setpfc() 411 struct hclge_dev *hdev = vport->back; in hclge_getdcbx() 423 struct hclge_dev *hdev = vport->back; in hclge_setdcbx() 442 struct hclge_dev *hdev = vport->back; in hclge_setup_tc()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/iavf/ |
D | iavf_client.c | 42 params->link_up = vsi->back->link_up; in iavf_client_get_params() 65 cinst = vsi->back->cinst; in iavf_notify_client_message() 68 dev_dbg(&vsi->back->pdev->dev, in iavf_notify_client_message() 90 cinst = vsi->back->cinst; in iavf_notify_client_l2_params() 94 dev_dbg(&vsi->back->pdev->dev, in iavf_notify_client_l2_params() 112 struct iavf_adapter *adapter = vsi->back; in iavf_notify_client_open() 118 dev_dbg(&vsi->back->pdev->dev, in iavf_notify_client_open() 164 struct iavf_adapter *adapter = vsi->back; in iavf_notify_client_close() 169 dev_dbg(&vsi->back->pdev->dev, in iavf_notify_client_close()
|
/kernel/linux/linux-5.10/Documentation/ABI/testing/ |
D | sysfs-class-bdi | 35 total write-back cache that relates to its current average 39 percentage of the write-back cache to a particular device. 45 given percentage of the write-back cache. This is useful in 47 most of the write-back cache. For example in case of an NFS
|
D | sysfs-platform-hidma-mgmt | 63 read transactions that can be issued back to back. 84 write transactions that can be issued back to back.
|
/kernel/linux/linux-5.10/arch/mips/include/asm/octeon/ |
D | cvmx-packet.h | 52 uint64_t back:4; member 63 uint64_t back:4;
|
D | cvmx-helper-util.h | 146 start_of_buffer = ((buffer_ptr.s.addr >> 7) - buffer_ptr.s.back) << 7; in cvmx_helper_free_packet_data() 160 ((buffer_ptr.s.addr >> 7) - buffer_ptr.s.back) << 7; in cvmx_helper_free_packet_data()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/i40e/ |
D | i40e_main.c | 129 struct i40e_pf *pf = (struct i40e_pf *)hw->back; in i40e_allocate_dma_mem_d() 147 struct i40e_pf *pf = (struct i40e_pf *)hw->back; in i40e_free_dma_mem_d() 326 struct i40e_pf *pf = vsi->back; in i40e_tx_timeout() 700 struct i40e_pf *pf = vsi->back; in i40e_update_eth_stats() 850 struct i40e_pf *pf = vsi->back; in i40e_update_vsi_stats() 1206 struct i40e_pf *pf = vsi->back; in i40e_update_stats() 1439 struct i40e_pf *pf = vsi->back; in i40e_rm_default_mac_filter() 1502 set_bit(__I40E_MACVLAN_SYNC_PENDING, vsi->back->state); in i40e_add_filter() 1552 set_bit(__I40E_MACVLAN_SYNC_PENDING, vsi->back->state); in __i40e_del_filter() 1657 struct i40e_pf *pf = vsi->back; in i40e_set_mac() [all …]
|
D | i40e_client.c | 49 struct i40e_dcbx_config *dcb_cfg = &vsi->back->hw.local_dcbx_config; in i40e_client_get_params() 64 dev_err(&vsi->back->pdev->dev, "Invalid queue set handle for TC = %d, vsi id = %d\n", in i40e_client_get_params() 86 struct i40e_pf *pf = vsi->back; in i40e_notify_client_of_vf_msg() 112 struct i40e_pf *pf = vsi->back; in i40e_notify_client_of_l2_param_changes() 119 dev_dbg(&vsi->back->pdev->dev, in i40e_notify_client_of_l2_param_changes() 124 dev_dbg(&vsi->back->pdev->dev, "Client is not open, abort l2 param change\n"); in i40e_notify_client_of_l2_param_changes() 171 struct i40e_pf *pf = vsi->back; in i40e_notify_client_of_netdev_close() 177 dev_dbg(&vsi->back->pdev->dev, in i40e_notify_client_of_netdev_close() 726 err = i40e_aq_update_vsi_params(&vsi->back->hw, &ctxt, NULL); in i40e_client_update_vsi_ctxt()
|
D | i40e_ethtool.c | 1061 struct i40e_pf *pf = np->vsi->back; in i40e_get_link_ksettings() 1153 struct i40e_pf *pf = np->vsi->back; in i40e_set_link_ksettings() 1388 struct i40e_pf *pf = np->vsi->back; in i40e_set_fec_cfg() 1450 struct i40e_pf *pf = np->vsi->back; in i40e_get_fec_param() 1492 struct i40e_pf *pf = np->vsi->back; in i40e_set_fec_param() 1536 struct i40e_pf *pf = np->vsi->back; in i40e_nway_reset() 1563 struct i40e_pf *pf = np->vsi->back; in i40e_get_pauseparam() 1598 struct i40e_pf *pf = np->vsi->back; in i40e_set_pauseparam() 1691 struct i40e_pf *pf = np->vsi->back; in i40e_get_msglevel() 1703 struct i40e_pf *pf = np->vsi->back; in i40e_set_msglevel() [all …]
|
/kernel/linux/linux-5.10/fs/xfs/libxfs/ |
D | xfs_da_btree.c | 136 to->back = be32_to_cpu(from3->hdr.info.hdr.back); in xfs_da3_node_hdr_from_disk() 144 to->back = be32_to_cpu(from->hdr.info.back); in xfs_da3_node_hdr_from_disk() 164 to3->hdr.info.hdr.back = cpu_to_be32(from->back); in xfs_da3_node_hdr_to_disk() 171 to->hdr.info.back = cpu_to_be32(from->back); in xfs_da3_node_hdr_to_disk() 604 node->hdr.info.back = cpu_to_be32(oldblk->blkno); in xfs_da3_split() 610 if (node->hdr.info.back) { in xfs_da3_split() 611 if (be32_to_cpu(node->hdr.info.back) != addblk->blkno) { in xfs_da3_split() 1160 ASSERT(!blkinfo->back); in xfs_da_blkinfo_onlychild_validate() 1191 ASSERT(oldroothdr.back == 0); in xfs_da3_root_join() 1310 forward = nodehdr.forw < nodehdr.back; in xfs_da3_node_toosmall() [all …]
|
/kernel/linux/linux-5.10/drivers/block/zram/zram_group/ |
D | Kconfig | 21 bool "Write back grouped zram objs to Hyperhold driver" 24 Write back grouped zram objs to hyperhold.
|
/kernel/linux/linux-5.10/Documentation/admin-guide/ |
D | dell_rbu.rst | 36 maintains a link list of packets for reading them back. 74 packets of data arranged back to back. It can be done as follows: 79 image file and then arrange all these packets back to back in to one single 120 read back the image downloaded.
|