Home
last modified time | relevance | path

Searched refs:qid (Results 1 – 25 of 248) sorted by relevance

12345678910

/kernel/linux/linux-5.10/arch/s390/include/asm/
Dap.h78 static inline struct ap_queue_status ap_tapq(ap_qid_t qid, unsigned long *info) in ap_tapq() argument
80 register unsigned long reg0 asm ("0") = qid; in ap_tapq()
101 static inline struct ap_queue_status ap_test_queue(ap_qid_t qid, in ap_test_queue() argument
106 qid |= 1UL << 23; /* set T bit*/ in ap_test_queue()
107 return ap_tapq(qid, info); in ap_test_queue()
116 static inline struct ap_queue_status ap_rapq(ap_qid_t qid) in ap_rapq() argument
118 register unsigned long reg0 asm ("0") = qid | (1UL << 24); in ap_rapq()
135 static inline struct ap_queue_status ap_zapq(ap_qid_t qid) in ap_zapq() argument
137 register unsigned long reg0 asm ("0") = qid | (2UL << 24); in ap_zapq()
219 static inline struct ap_queue_status ap_aqic(ap_qid_t qid, in ap_aqic() argument
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/hw/cxgb4/
Dresource.c111 u32 qid; in c4iw_get_cqid() local
119 qid = entry->qid; in c4iw_get_cqid()
122 qid = c4iw_get_resource(&rdev->resource.qid_table); in c4iw_get_cqid()
123 if (!qid) in c4iw_get_cqid()
126 rdev->stats.qid.cur += rdev->qpmask + 1; in c4iw_get_cqid()
128 for (i = qid+1; i & rdev->qpmask; i++) { in c4iw_get_cqid()
132 entry->qid = i; in c4iw_get_cqid()
143 entry->qid = qid; in c4iw_get_cqid()
145 for (i = qid+1; i & rdev->qpmask; i++) { in c4iw_get_cqid()
149 entry->qid = i; in c4iw_get_cqid()
[all …]
/kernel/linux/linux-5.10/drivers/scsi/lpfc/
Dlpfc_debugfs.h561 lpfc_debug_dump_wq_by_id(struct lpfc_hba *phba, int qid) in lpfc_debug_dump_wq_by_id() argument
566 if (phba->sli4_hba.hdwq[wq_idx].io_wq->queue_id == qid) in lpfc_debug_dump_wq_by_id()
569 pr_err("IO WQ[Idx:%d|Qid:%d]\n", wq_idx, qid); in lpfc_debug_dump_wq_by_id()
574 if (phba->sli4_hba.els_wq->queue_id == qid) { in lpfc_debug_dump_wq_by_id()
575 pr_err("ELS WQ[Qid:%d]\n", qid); in lpfc_debug_dump_wq_by_id()
580 if (phba->sli4_hba.nvmels_wq->queue_id == qid) { in lpfc_debug_dump_wq_by_id()
581 pr_err("NVME LS WQ[Qid:%d]\n", qid); in lpfc_debug_dump_wq_by_id()
595 lpfc_debug_dump_mq_by_id(struct lpfc_hba *phba, int qid) in lpfc_debug_dump_mq_by_id() argument
597 if (phba->sli4_hba.mbx_wq->queue_id == qid) { in lpfc_debug_dump_mq_by_id()
598 printk(KERN_ERR "MBX WQ[Qid:%d]\n", qid); in lpfc_debug_dump_mq_by_id()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdkfd/
Dkfd_process_queue_manager.c32 struct process_queue_manager *pqm, unsigned int qid) in get_queue_by_qid() argument
37 if ((pqn->q && pqn->q->properties.queue_id == qid) || in get_queue_by_qid()
38 (pqn->kq && pqn->kq->queue->properties.queue_id == qid)) in get_queue_by_qid()
46 unsigned int *qid) in find_available_queue_slot() argument
62 *qid = found; in find_available_queue_slot()
78 int pqm_set_gws(struct process_queue_manager *pqm, unsigned int qid, in pqm_set_gws() argument
87 pqn = get_queue_by_qid(pqm, qid); in pqm_set_gws()
169 struct file *f, unsigned int qid) in init_user_queue() argument
178 q_properties->queue_id = qid; in init_user_queue()
196 unsigned int *qid, in pqm_create_queue() argument
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/prestera/
Dprestera_pci.c197 static u32 prestera_fw_evtq_len(struct prestera_fw *fw, u8 qid) in prestera_fw_evtq_len() argument
199 return fw->evt_queue[qid].len; in prestera_fw_evtq_len()
202 static u32 prestera_fw_evtq_avail(struct prestera_fw *fw, u8 qid) in prestera_fw_evtq_avail() argument
204 u32 wr_idx = prestera_fw_read(fw, PRESTERA_EVTQ_WR_IDX_REG(qid)); in prestera_fw_evtq_avail()
205 u32 rd_idx = prestera_fw_read(fw, PRESTERA_EVTQ_RD_IDX_REG(qid)); in prestera_fw_evtq_avail()
207 return CIRC_CNT(wr_idx, rd_idx, prestera_fw_evtq_len(fw, qid)); in prestera_fw_evtq_avail()
211 u8 qid, u32 idx) in prestera_fw_evtq_rd_set() argument
213 u32 rd_idx = idx & (prestera_fw_evtq_len(fw, qid) - 1); in prestera_fw_evtq_rd_set()
215 prestera_fw_write(fw, PRESTERA_EVTQ_RD_IDX_REG(qid), rd_idx); in prestera_fw_evtq_rd_set()
218 static u8 __iomem *prestera_fw_evtq_buf(struct prestera_fw *fw, u8 qid) in prestera_fw_evtq_buf() argument
[all …]
/kernel/linux/linux-5.10/drivers/s390/crypto/
Dap_queue.c37 status = ap_aqic(aq->qid, qirqctrl, ind); in ap_queue_enable_irq()
47 AP_QID_CARD(aq->qid), in ap_queue_enable_irq()
48 AP_QID_QUEUE(aq->qid)); in ap_queue_enable_irq()
71 __ap_send(ap_qid_t qid, unsigned long long psmid, void *msg, size_t length, in __ap_send() argument
75 qid |= 0x400000UL; in __ap_send()
76 return ap_nqap(qid, psmid, msg, length); in __ap_send()
79 int ap_send(ap_qid_t qid, unsigned long long psmid, void *msg, size_t length) in ap_send() argument
83 status = __ap_send(qid, psmid, msg, length, 0); in ap_send()
98 int ap_recv(ap_qid_t qid, unsigned long long *psmid, void *msg, size_t length) in ap_recv() argument
104 status = ap_dqap(qid, psmid, msg, length); in ap_recv()
[all …]
Dzcrypt_api.c642 int cpen, qpen, qid = 0, rc = -ENODEV; in zcrypt_rsa_modexpo() local
701 AP_QID_QUEUE(zq->queue->qid))) in zcrypt_rsa_modexpo()
705 tr->last_qid == zq->queue->qid) ? in zcrypt_rsa_modexpo()
723 qid = pref_zq->queue->qid; in zcrypt_rsa_modexpo()
734 tr->last_qid = qid; in zcrypt_rsa_modexpo()
737 AP_QID_CARD(qid), AP_QID_QUEUE(qid)); in zcrypt_rsa_modexpo()
750 int cpen, qpen, qid = 0, rc = -ENODEV; in zcrypt_rsa_crt() local
809 AP_QID_QUEUE(zq->queue->qid))) in zcrypt_rsa_crt()
813 tr->last_qid == zq->queue->qid) ? in zcrypt_rsa_crt()
831 qid = pref_zq->queue->qid; in zcrypt_rsa_crt()
[all …]
Dap_bus.c300 static bool ap_queue_info(ap_qid_t qid, int *q_type, in ap_queue_info() argument
307 if (AP_QID_CARD(qid) > ap_max_adapter_id || in ap_queue_info()
308 AP_QID_QUEUE(qid) > ap_max_domain_id) in ap_queue_info()
312 status = ap_test_queue(qid, ap_apft_available(), &info); in ap_queue_info()
591 AP_QID_CARD(to_ap_queue(dev)->qid) == (int)(long) data) in __ap_queue_devices_with_id_unregister()
607 card = AP_QID_CARD(to_ap_queue(dev)->qid); in __ap_revise_reserved()
608 queue = AP_QID_QUEUE(to_ap_queue(dev)->qid); in __ap_revise_reserved()
686 card = AP_QID_CARD(to_ap_queue(dev)->qid); in ap_device_probe()
687 queue = AP_QID_QUEUE(to_ap_queue(dev)->qid); in ap_device_probe()
701 to_ap_queue(dev)->qid); in ap_device_probe()
[all …]
Dzcrypt_msgtype6.c297 msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid); in ICAMEX_msg_to_type6MEX_msgX()
367 msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid); in ICACRT_msg_to_type6CRT_msgX()
664 AP_QID_CARD(zq->queue->qid), in convert_type86_ica()
665 AP_QID_QUEUE(zq->queue->qid), in convert_type86_ica()
671 AP_QID_CARD(zq->queue->qid), in convert_type86_ica()
672 AP_QID_QUEUE(zq->queue->qid), in convert_type86_ica()
675 AP_QID_CARD(zq->queue->qid), in convert_type86_ica()
676 AP_QID_QUEUE(zq->queue->qid), in convert_type86_ica()
816 AP_QID_CARD(zq->queue->qid), in convert_response_ica()
817 AP_QID_QUEUE(zq->queue->qid), in convert_response_ica()
[all …]
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt76/mt7615/
Dsdio_txrx.c83 static int mt7663s_rx_run_queue(struct mt76_dev *dev, enum mt76_rxq_id qid, in mt7663s_rx_run_queue() argument
86 struct mt76_queue *q = &dev->q_rx[qid]; in mt7663s_rx_run_queue()
92 for (i = 0; i < intr->rx.num[qid]; i++) in mt7663s_rx_run_queue()
93 len += round_up(intr->rx.len[qid][i] + 4, 4); in mt7663s_rx_run_queue()
108 err = sdio_readsb(sdio->func, buf, MCR_WRDR(qid), len); in mt7663s_rx_run_queue()
117 for (i = 0; i < intr->rx.num[qid]; i++) { in mt7663s_rx_run_queue()
121 len = intr->rx.len[qid][i]; in mt7663s_rx_run_queue()
140 static int mt7663s_tx_pick_quota(struct mt76_sdio *sdio, enum mt76_txq_id qid, in mt7663s_tx_pick_quota() argument
147 if (qid == MT_TXQ_MCU) { in mt7663s_tx_pick_quota()
162 static void mt7663s_tx_update_quota(struct mt76_sdio *sdio, enum mt76_txq_id qid, in mt7663s_tx_update_quota() argument
[all …]
/kernel/linux/linux-5.10/drivers/nvme/host/
Dtrace.h26 #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ argument
29 ((qid) ? \
53 __field(int, qid)
64 __entry->qid = nvme_req_qid(req);
77 __entry->qid, __entry->cid, __entry->nsid,
79 show_opcode_name(__entry->qid, __entry->opcode,
81 parse_nvme_cmd(__entry->qid, __entry->opcode,
91 __field(int, qid)
100 __entry->qid = nvme_req_qid(req);
110 __entry->qid, __entry->cid, __entry->result,
[all …]
/kernel/linux/linux-5.10/drivers/nvme/target/
Dtrace.h28 #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ argument
31 (qid ? \
65 __field(int, qid)
78 __entry->qid = req->sq->qid;
92 __entry->qid, __entry->cid, __entry->nsid,
94 show_opcode_name(__entry->qid, __entry->opcode,
96 parse_nvme_cmd(__entry->qid, __entry->opcode,
106 __field(int, qid)
113 __entry->qid = req->cq->qid;
122 __entry->qid, __entry->cid, __entry->result, __entry->status)
Dfabrics-cmd.c109 u16 qid = le16_to_cpu(c->qid); in nvmet_install_queue() local
129 nvmet_cq_setup(ctrl, req->cq, qid, sqsize + 1); in nvmet_install_queue()
130 nvmet_sq_setup(ctrl, req->sq, qid, sqsize + 1); in nvmet_install_queue()
141 qid, ctrl->cntlid, ret); in nvmet_install_queue()
227 u16 qid = le16_to_cpu(c->qid); in nvmet_execute_io_connect() local
259 if (unlikely(qid > ctrl->subsys->max_qid)) { in nvmet_execute_io_connect()
260 pr_warn("invalid queue id (%d)\n", qid); in nvmet_execute_io_connect()
262 req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(qid); in nvmet_execute_io_connect()
273 pr_debug("adding queue %d to ctrl %d.\n", qid, ctrl->cntlid); in nvmet_execute_io_connect()
303 if (cmd->connect.qid == 0) in nvmet_parse_connect_cmd()
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ixgbe/
Dixgbe_xsk.c15 int qid = ring->ring_idx; in ixgbe_xsk_pool() local
17 if (!xdp_on || !test_bit(qid, adapter->af_xdp_zc_qps)) in ixgbe_xsk_pool()
20 return xsk_get_pool_from_qid(adapter->netdev, qid); in ixgbe_xsk_pool()
25 u16 qid) in ixgbe_xsk_pool_enable() argument
31 if (qid >= adapter->num_rx_queues) in ixgbe_xsk_pool_enable()
34 if (qid >= netdev->real_num_rx_queues || in ixgbe_xsk_pool_enable()
35 qid >= netdev->real_num_tx_queues) in ixgbe_xsk_pool_enable()
46 ixgbe_txrx_ring_disable(adapter, qid); in ixgbe_xsk_pool_enable()
48 set_bit(qid, adapter->af_xdp_zc_qps); in ixgbe_xsk_pool_enable()
51 ixgbe_txrx_ring_enable(adapter, qid); in ixgbe_xsk_pool_enable()
[all …]
/kernel/linux/linux-5.10/drivers/vdpa/ifcvf/
Difcvf_main.c245 static int ifcvf_vdpa_get_vq_state(struct vdpa_device *vdpa_dev, u16 qid, in ifcvf_vdpa_get_vq_state() argument
250 state->avail_index = ifcvf_get_vq_state(vf, qid); in ifcvf_vdpa_get_vq_state()
254 static int ifcvf_vdpa_set_vq_state(struct vdpa_device *vdpa_dev, u16 qid, in ifcvf_vdpa_set_vq_state() argument
259 return ifcvf_set_vq_state(vf, qid, state->avail_index); in ifcvf_vdpa_set_vq_state()
262 static void ifcvf_vdpa_set_vq_cb(struct vdpa_device *vdpa_dev, u16 qid, in ifcvf_vdpa_set_vq_cb() argument
267 vf->vring[qid].cb = *cb; in ifcvf_vdpa_set_vq_cb()
271 u16 qid, bool ready) in ifcvf_vdpa_set_vq_ready() argument
275 vf->vring[qid].ready = ready; in ifcvf_vdpa_set_vq_ready()
278 static bool ifcvf_vdpa_get_vq_ready(struct vdpa_device *vdpa_dev, u16 qid) in ifcvf_vdpa_get_vq_ready() argument
282 return vf->vring[qid].ready; in ifcvf_vdpa_get_vq_ready()
[all …]
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt76/
Dsdio.c21 mt76s_alloc_rx_queue(struct mt76_dev *dev, enum mt76_rxq_id qid) in mt76s_alloc_rx_queue() argument
23 struct mt76_queue *q = &dev->q_rx[qid]; in mt76s_alloc_rx_queue()
111 int qid = q - &dev->q_rx[MT_RXQ_MAIN]; in mt76s_process_rx_queue() local
128 if (qid == MT_RXQ_MAIN) in mt76s_process_rx_queue()
134 static void mt76s_process_tx_queue(struct mt76_dev *dev, enum mt76_txq_id qid) in mt76s_process_tx_queue() argument
136 struct mt76_queue *q = dev->q_tx[qid]; in mt76s_process_tx_queue()
147 if (qid == MT_TXQ_MCU) { in mt76s_process_tx_queue()
162 if (qid == MT_TXQ_MCU) in mt76s_process_tx_queue()
165 mt76_txq_schedule(&dev->phy, qid); in mt76s_process_tx_queue()
168 ieee80211_wake_queue(dev->hw, qid); in mt76s_process_tx_queue()
[all …]
Dtx.c228 __mt76_tx_queue_skb(struct mt76_dev *dev, int qid, struct sk_buff *skb, in __mt76_tx_queue_skb() argument
239 idx = dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, sta); in __mt76_tx_queue_skb()
244 q = dev->q_tx[qid]; in __mt76_tx_queue_skb()
261 int qid = skb_get_queue_mapping(skb); in mt76_tx() local
269 if (WARN_ON(qid >= MT_TXQ_PSD)) { in mt76_tx()
270 qid = MT_TXQ_BE; in mt76_tx()
271 skb_set_queue_mapping(skb, qid); in mt76_tx()
277 qid = MT_TXQ_PSD; in mt76_tx()
278 skb_set_queue_mapping(skb, qid); in mt76_tx()
288 q = dev->q_tx[qid]; in mt76_tx()
[all …]
/kernel/linux/linux-5.10/fs/9p/
Dcache.c86 if (buflen != sizeof(v9inode->qid.version)) in v9fs_cache_inode_check_aux()
89 if (memcmp(buffer, &v9inode->qid.version, in v9fs_cache_inode_check_aux()
90 sizeof(v9inode->qid.version))) in v9fs_cache_inode_check_aux()
117 &v9inode->qid.path, in v9fs_cache_inode_get_cookie()
118 sizeof(v9inode->qid.path), in v9fs_cache_inode_get_cookie()
119 &v9inode->qid.version, in v9fs_cache_inode_get_cookie()
120 sizeof(v9inode->qid.version), in v9fs_cache_inode_get_cookie()
138 fscache_relinquish_cookie(v9inode->fscache, &v9inode->qid.version, in v9fs_cache_inode_put_cookie()
190 &v9inode->qid.path, in v9fs_cache_inode_reset_cookie()
191 sizeof(v9inode->qid.path), in v9fs_cache_inode_reset_cookie()
[all …]
/kernel/linux/linux-5.10/include/linux/
Dquota.h79 extern qid_t from_kqid(struct user_namespace *to, struct kqid qid);
80 extern qid_t from_kqid_munged(struct user_namespace *to, struct kqid qid);
81 extern bool qid_valid(struct kqid qid);
98 enum quota_type type, qid_t qid) in make_kqid() argument
105 kqid.uid = make_kuid(from, qid); in make_kqid()
108 kqid.gid = make_kgid(from, qid); in make_kqid()
111 kqid.projid = make_kprojid(from, qid); in make_kqid()
187 static inline bool qid_has_mapping(struct user_namespace *ns, struct kqid qid) in qid_has_mapping() argument
189 return from_kqid(ns, qid) != (qid_t) -1; in qid_has_mapping()
318 …int (*get_next_id)(struct super_block *sb, struct kqid *qid); /* Get next ID with existing structu…
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/en/
Dparams.h44 u16 qid, in mlx5e_qid_get_ch_if_in_group() argument
49 int ch = qid - nch * group; in mlx5e_qid_get_ch_if_in_group()
59 u16 qid, in mlx5e_qid_get_ch_and_group() argument
65 *ix = qid % nch; in mlx5e_qid_get_ch_and_group()
66 *group = qid / nch; in mlx5e_qid_get_ch_and_group()
70 struct mlx5e_params *params, u64 qid) in mlx5e_qid_validate() argument
72 return qid < params->num_channels * profile->rq_groups; in mlx5e_qid_validate()
/kernel/linux/linux-5.10/fs/xfs/
Dxfs_quotaops.c220 struct kqid qid, in xfs_fs_get_dqblk() argument
231 id = from_kqid(&init_user_ns, qid); in xfs_fs_get_dqblk()
232 return xfs_qm_scall_getquota(mp, id, xfs_quota_type(qid.type), qdq); in xfs_fs_get_dqblk()
239 struct kqid *qid, in xfs_fs_get_nextdqblk() argument
251 id = from_kqid(&init_user_ns, *qid); in xfs_fs_get_nextdqblk()
252 ret = xfs_qm_scall_getquota_next(mp, &id, xfs_quota_type(qid->type), in xfs_fs_get_nextdqblk()
258 *qid = make_kqid(current_user_ns(), qid->type, id); in xfs_fs_get_nextdqblk()
265 struct kqid qid, in xfs_fs_set_dqblk() argument
277 return xfs_qm_scall_setqlim(mp, from_kqid(&init_user_ns, qid), in xfs_fs_set_dqblk()
278 xfs_quota_type(qid.type), qdq); in xfs_fs_set_dqblk()
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/i40e/
Di40e_xsk.c42 u16 qid) in i40e_xsk_pool_enable() argument
51 if (qid >= vsi->num_queue_pairs) in i40e_xsk_pool_enable()
54 if (qid >= netdev->real_num_rx_queues || in i40e_xsk_pool_enable()
55 qid >= netdev->real_num_tx_queues) in i40e_xsk_pool_enable()
62 set_bit(qid, vsi->af_xdp_zc_qps); in i40e_xsk_pool_enable()
67 err = i40e_queue_pair_disable(vsi, qid); in i40e_xsk_pool_enable()
71 err = i40e_queue_pair_enable(vsi, qid); in i40e_xsk_pool_enable()
76 err = i40e_xsk_wakeup(vsi->netdev, qid, XDP_WAKEUP_RX); in i40e_xsk_pool_enable()
92 static int i40e_xsk_pool_disable(struct i40e_vsi *vsi, u16 qid) in i40e_xsk_pool_disable() argument
99 pool = xsk_get_pool_from_qid(netdev, qid); in i40e_xsk_pool_disable()
[all …]
/kernel/linux/linux-5.10/fs/quota/
Dkqid.c120 bool qid_valid(struct kqid qid) in qid_valid() argument
122 switch (qid.type) { in qid_valid()
124 return uid_valid(qid.uid); in qid_valid()
126 return gid_valid(qid.gid); in qid_valid()
128 return projid_valid(qid.projid); in qid_valid()
Dquota.c201 struct kqid qid; in quota_getquota() local
208 qid = make_kqid(current_user_ns(), type, id); in quota_getquota()
209 if (!qid_has_mapping(sb->s_user_ns, qid)) in quota_getquota()
211 ret = sb->s_qcop->get_dqblk(sb, qid, &fdq); in quota_getquota()
237 struct kqid qid; in quota_getnextquota() local
244 qid = make_kqid(current_user_ns(), type, id); in quota_getnextquota()
245 if (!qid_has_mapping(sb->s_user_ns, qid)) in quota_getnextquota()
247 ret = sb->s_qcop->get_nextdqblk(sb, &qid, &fdq); in quota_getnextquota()
252 idq.dqb_id = from_kqid(current_user_ns(), qid); in quota_getnextquota()
289 struct kqid qid; in quota_setquota() local
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ice/
Dice_xsk.c289 static void ice_xsk_remove_pool(struct ice_vsi *vsi, u16 qid) in ice_xsk_remove_pool() argument
291 vsi->xsk_pools[qid] = NULL; in ice_xsk_remove_pool()
308 static int ice_xsk_pool_disable(struct ice_vsi *vsi, u16 qid) in ice_xsk_pool_disable() argument
310 if (!vsi->xsk_pools || qid >= vsi->num_xsk_pools || in ice_xsk_pool_disable()
311 !vsi->xsk_pools[qid]) in ice_xsk_pool_disable()
314 xsk_pool_dma_unmap(vsi->xsk_pools[qid], ICE_RX_DMA_ATTR); in ice_xsk_pool_disable()
315 ice_xsk_remove_pool(vsi, qid); in ice_xsk_pool_disable()
329 ice_xsk_pool_enable(struct ice_vsi *vsi, struct xsk_buff_pool *pool, u16 qid) in ice_xsk_pool_enable() argument
338 if (qid >= vsi->num_xsk_pools) in ice_xsk_pool_enable()
345 if (vsi->xsk_pools && vsi->xsk_pools[qid]) in ice_xsk_pool_enable()
[all …]

12345678910