/drivers/infiniband/hw/cxgb4/ |
D | resource.c | 111 u32 qid; in c4iw_get_cqid() local 119 qid = entry->qid; in c4iw_get_cqid() 122 qid = c4iw_get_resource(&rdev->resource.qid_table); in c4iw_get_cqid() 123 if (!qid) in c4iw_get_cqid() 126 rdev->stats.qid.cur += rdev->qpmask + 1; in c4iw_get_cqid() 128 for (i = qid+1; i & rdev->qpmask; i++) { in c4iw_get_cqid() 132 entry->qid = i; in c4iw_get_cqid() 143 entry->qid = qid; in c4iw_get_cqid() 145 for (i = qid+1; i & rdev->qpmask; i++) { in c4iw_get_cqid() 149 entry->qid = i; in c4iw_get_cqid() [all …]
|
/drivers/scsi/lpfc/ |
D | lpfc_debugfs.h | 561 lpfc_debug_dump_wq_by_id(struct lpfc_hba *phba, int qid) in lpfc_debug_dump_wq_by_id() argument 566 if (phba->sli4_hba.hdwq[wq_idx].io_wq->queue_id == qid) in lpfc_debug_dump_wq_by_id() 569 pr_err("IO WQ[Idx:%d|Qid:%d]\n", wq_idx, qid); in lpfc_debug_dump_wq_by_id() 574 if (phba->sli4_hba.els_wq->queue_id == qid) { in lpfc_debug_dump_wq_by_id() 575 pr_err("ELS WQ[Qid:%d]\n", qid); in lpfc_debug_dump_wq_by_id() 580 if (phba->sli4_hba.nvmels_wq->queue_id == qid) { in lpfc_debug_dump_wq_by_id() 581 pr_err("NVME LS WQ[Qid:%d]\n", qid); in lpfc_debug_dump_wq_by_id() 595 lpfc_debug_dump_mq_by_id(struct lpfc_hba *phba, int qid) in lpfc_debug_dump_mq_by_id() argument 597 if (phba->sli4_hba.mbx_wq->queue_id == qid) { in lpfc_debug_dump_mq_by_id() 598 printk(KERN_ERR "MBX WQ[Qid:%d]\n", qid); in lpfc_debug_dump_mq_by_id() [all …]
|
/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_process_queue_manager.c | 32 struct process_queue_manager *pqm, unsigned int qid) in get_queue_by_qid() argument 37 if ((pqn->q && pqn->q->properties.queue_id == qid) || in get_queue_by_qid() 38 (pqn->kq && pqn->kq->queue->properties.queue_id == qid)) in get_queue_by_qid() 46 unsigned int *qid) in find_available_queue_slot() argument 62 *qid = found; in find_available_queue_slot() 78 int pqm_set_gws(struct process_queue_manager *pqm, unsigned int qid, in pqm_set_gws() argument 87 pqn = get_queue_by_qid(pqm, qid); in pqm_set_gws() 169 struct file *f, unsigned int qid) in init_user_queue() argument 178 q_properties->queue_id = qid; in init_user_queue() 196 unsigned int *qid, in pqm_create_queue() argument [all …]
|
/drivers/net/ethernet/marvell/prestera/ |
D | prestera_pci.c | 197 static u32 prestera_fw_evtq_len(struct prestera_fw *fw, u8 qid) in prestera_fw_evtq_len() argument 199 return fw->evt_queue[qid].len; in prestera_fw_evtq_len() 202 static u32 prestera_fw_evtq_avail(struct prestera_fw *fw, u8 qid) in prestera_fw_evtq_avail() argument 204 u32 wr_idx = prestera_fw_read(fw, PRESTERA_EVTQ_WR_IDX_REG(qid)); in prestera_fw_evtq_avail() 205 u32 rd_idx = prestera_fw_read(fw, PRESTERA_EVTQ_RD_IDX_REG(qid)); in prestera_fw_evtq_avail() 207 return CIRC_CNT(wr_idx, rd_idx, prestera_fw_evtq_len(fw, qid)); in prestera_fw_evtq_avail() 211 u8 qid, u32 idx) in prestera_fw_evtq_rd_set() argument 213 u32 rd_idx = idx & (prestera_fw_evtq_len(fw, qid) - 1); in prestera_fw_evtq_rd_set() 215 prestera_fw_write(fw, PRESTERA_EVTQ_RD_IDX_REG(qid), rd_idx); in prestera_fw_evtq_rd_set() 218 static u8 __iomem *prestera_fw_evtq_buf(struct prestera_fw *fw, u8 qid) in prestera_fw_evtq_buf() argument [all …]
|
/drivers/s390/crypto/ |
D | ap_queue.c | 37 status = ap_aqic(aq->qid, qirqctrl, ind); in ap_queue_enable_irq() 47 AP_QID_CARD(aq->qid), in ap_queue_enable_irq() 48 AP_QID_QUEUE(aq->qid)); in ap_queue_enable_irq() 71 __ap_send(ap_qid_t qid, unsigned long long psmid, void *msg, size_t length, in __ap_send() argument 75 qid |= 0x400000UL; in __ap_send() 76 return ap_nqap(qid, psmid, msg, length); in __ap_send() 79 int ap_send(ap_qid_t qid, unsigned long long psmid, void *msg, size_t length) in ap_send() argument 83 status = __ap_send(qid, psmid, msg, length, 0); in ap_send() 98 int ap_recv(ap_qid_t qid, unsigned long long *psmid, void *msg, size_t length) in ap_recv() argument 104 status = ap_dqap(qid, psmid, msg, length); in ap_recv() [all …]
|
D | zcrypt_api.c | 643 int cpen, qpen, qid = 0, rc = -ENODEV; in zcrypt_rsa_modexpo() local 702 AP_QID_QUEUE(zq->queue->qid))) in zcrypt_rsa_modexpo() 706 tr->last_qid == zq->queue->qid) ? in zcrypt_rsa_modexpo() 724 qid = pref_zq->queue->qid; in zcrypt_rsa_modexpo() 735 tr->last_qid = qid; in zcrypt_rsa_modexpo() 738 AP_QID_CARD(qid), AP_QID_QUEUE(qid)); in zcrypt_rsa_modexpo() 751 int cpen, qpen, qid = 0, rc = -ENODEV; in zcrypt_rsa_crt() local 810 AP_QID_QUEUE(zq->queue->qid))) in zcrypt_rsa_crt() 814 tr->last_qid == zq->queue->qid) ? in zcrypt_rsa_crt() 832 qid = pref_zq->queue->qid; in zcrypt_rsa_crt() [all …]
|
D | ap_bus.c | 300 static bool ap_queue_info(ap_qid_t qid, int *q_type, in ap_queue_info() argument 307 if (AP_QID_CARD(qid) > ap_max_adapter_id || in ap_queue_info() 308 AP_QID_QUEUE(qid) > ap_max_domain_id) in ap_queue_info() 312 status = ap_test_queue(qid, ap_apft_available(), &info); in ap_queue_info() 591 AP_QID_CARD(to_ap_queue(dev)->qid) == (int)(long) data) in __ap_queue_devices_with_id_unregister() 607 card = AP_QID_CARD(to_ap_queue(dev)->qid); in __ap_revise_reserved() 608 queue = AP_QID_QUEUE(to_ap_queue(dev)->qid); in __ap_revise_reserved() 686 card = AP_QID_CARD(to_ap_queue(dev)->qid); in ap_device_probe() 687 queue = AP_QID_QUEUE(to_ap_queue(dev)->qid); in ap_device_probe() 701 to_ap_queue(dev)->qid); in ap_device_probe() [all …]
|
D | zcrypt_msgtype6.c | 297 msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid); in ICAMEX_msg_to_type6MEX_msgX() 367 msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid); in ICACRT_msg_to_type6CRT_msgX() 664 AP_QID_CARD(zq->queue->qid), in convert_type86_ica() 665 AP_QID_QUEUE(zq->queue->qid), in convert_type86_ica() 671 AP_QID_CARD(zq->queue->qid), in convert_type86_ica() 672 AP_QID_QUEUE(zq->queue->qid), in convert_type86_ica() 675 AP_QID_CARD(zq->queue->qid), in convert_type86_ica() 676 AP_QID_QUEUE(zq->queue->qid), in convert_type86_ica() 816 AP_QID_CARD(zq->queue->qid), in convert_response_ica() 817 AP_QID_QUEUE(zq->queue->qid), in convert_response_ica() [all …]
|
D | zcrypt_queue.c | 69 AP_QID_CARD(zq->queue->qid), in online_store() 70 AP_QID_QUEUE(zq->queue->qid), in online_store() 173 AP_QID_CARD(zq->queue->qid), AP_QID_QUEUE(zq->queue->qid)); in zcrypt_queue_register() 214 AP_QID_CARD(zq->queue->qid), AP_QID_QUEUE(zq->queue->qid)); in zcrypt_queue_unregister()
|
/drivers/net/wireless/mediatek/mt76/mt7615/ |
D | sdio_txrx.c | 83 static int mt7663s_rx_run_queue(struct mt76_dev *dev, enum mt76_rxq_id qid, in mt7663s_rx_run_queue() argument 86 struct mt76_queue *q = &dev->q_rx[qid]; in mt7663s_rx_run_queue() 92 for (i = 0; i < intr->rx.num[qid]; i++) in mt7663s_rx_run_queue() 93 len += round_up(intr->rx.len[qid][i] + 4, 4); in mt7663s_rx_run_queue() 108 err = sdio_readsb(sdio->func, buf, MCR_WRDR(qid), len); in mt7663s_rx_run_queue() 117 for (i = 0; i < intr->rx.num[qid]; i++) { in mt7663s_rx_run_queue() 121 len = intr->rx.len[qid][i]; in mt7663s_rx_run_queue() 140 static int mt7663s_tx_pick_quota(struct mt76_sdio *sdio, enum mt76_txq_id qid, in mt7663s_tx_pick_quota() argument 147 if (qid == MT_TXQ_MCU) { in mt7663s_tx_pick_quota() 162 static void mt7663s_tx_update_quota(struct mt76_sdio *sdio, enum mt76_txq_id qid, in mt7663s_tx_update_quota() argument [all …]
|
/drivers/nvme/host/ |
D | trace.h | 26 #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ argument 29 ((qid) ? \ 53 __field(int, qid) 64 __entry->qid = nvme_req_qid(req); 77 __entry->qid, __entry->cid, __entry->nsid, 79 show_opcode_name(__entry->qid, __entry->opcode, 81 parse_nvme_cmd(__entry->qid, __entry->opcode, 91 __field(int, qid) 100 __entry->qid = nvme_req_qid(req); 110 __entry->qid, __entry->cid, __entry->result, [all …]
|
D | pci.c | 170 static inline unsigned int sq_idx(unsigned int qid, u32 stride) in sq_idx() argument 172 return qid * 2 * stride; in sq_idx() 175 static inline unsigned int cq_idx(unsigned int qid, u32 stride) in cq_idx() argument 177 return (qid * 2 + 1) * stride; in cq_idx() 204 u16 qid; member 286 struct nvme_queue *nvmeq, int qid) in nvme_dbbuf_init() argument 288 if (!dev->dbbuf_dbs || !qid) in nvme_dbbuf_init() 291 nvmeq->dbbuf_sq_db = &dev->dbbuf_dbs[sq_idx(qid, dev->db_stride)]; in nvme_dbbuf_init() 292 nvmeq->dbbuf_cq_db = &dev->dbbuf_dbs[cq_idx(qid, dev->db_stride)]; in nvme_dbbuf_init() 293 nvmeq->dbbuf_sq_ei = &dev->dbbuf_eis[sq_idx(qid, dev->db_stride)]; in nvme_dbbuf_init() [all …]
|
/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_xsk.c | 15 int qid = ring->ring_idx; in ixgbe_xsk_pool() local 17 if (!xdp_on || !test_bit(qid, adapter->af_xdp_zc_qps)) in ixgbe_xsk_pool() 20 return xsk_get_pool_from_qid(adapter->netdev, qid); in ixgbe_xsk_pool() 25 u16 qid) in ixgbe_xsk_pool_enable() argument 31 if (qid >= adapter->num_rx_queues) in ixgbe_xsk_pool_enable() 34 if (qid >= netdev->real_num_rx_queues || in ixgbe_xsk_pool_enable() 35 qid >= netdev->real_num_tx_queues) in ixgbe_xsk_pool_enable() 46 ixgbe_txrx_ring_disable(adapter, qid); in ixgbe_xsk_pool_enable() 48 set_bit(qid, adapter->af_xdp_zc_qps); in ixgbe_xsk_pool_enable() 51 ixgbe_txrx_ring_enable(adapter, qid); in ixgbe_xsk_pool_enable() [all …]
|
/drivers/nvme/target/ |
D | trace.h | 28 #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ argument 31 (qid ? \ 65 __field(int, qid) 78 __entry->qid = req->sq->qid; 92 __entry->qid, __entry->cid, __entry->nsid, 94 show_opcode_name(__entry->qid, __entry->opcode, 96 parse_nvme_cmd(__entry->qid, __entry->opcode, 106 __field(int, qid) 113 __entry->qid = req->cq->qid; 122 __entry->qid, __entry->cid, __entry->result, __entry->status)
|
D | fabrics-cmd.c | 109 u16 qid = le16_to_cpu(c->qid); in nvmet_install_queue() local 129 nvmet_cq_setup(ctrl, req->cq, qid, sqsize + 1); in nvmet_install_queue() 130 nvmet_sq_setup(ctrl, req->sq, qid, sqsize + 1); in nvmet_install_queue() 141 qid, ctrl->cntlid, ret); in nvmet_install_queue() 229 u16 qid = le16_to_cpu(c->qid); in nvmet_execute_io_connect() local 264 if (unlikely(qid > ctrl->subsys->max_qid)) { in nvmet_execute_io_connect() 265 pr_warn("invalid queue id (%d)\n", qid); in nvmet_execute_io_connect() 267 req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(qid); in nvmet_execute_io_connect() 278 pr_debug("adding queue %d to ctrl %d.\n", qid, ctrl->cntlid); in nvmet_execute_io_connect() 308 if (cmd->connect.qid == 0) in nvmet_parse_connect_cmd()
|
/drivers/vdpa/ifcvf/ |
D | ifcvf_main.c | 245 static int ifcvf_vdpa_get_vq_state(struct vdpa_device *vdpa_dev, u16 qid, in ifcvf_vdpa_get_vq_state() argument 250 state->avail_index = ifcvf_get_vq_state(vf, qid); in ifcvf_vdpa_get_vq_state() 254 static int ifcvf_vdpa_set_vq_state(struct vdpa_device *vdpa_dev, u16 qid, in ifcvf_vdpa_set_vq_state() argument 259 return ifcvf_set_vq_state(vf, qid, state->avail_index); in ifcvf_vdpa_set_vq_state() 262 static void ifcvf_vdpa_set_vq_cb(struct vdpa_device *vdpa_dev, u16 qid, in ifcvf_vdpa_set_vq_cb() argument 267 vf->vring[qid].cb = *cb; in ifcvf_vdpa_set_vq_cb() 271 u16 qid, bool ready) in ifcvf_vdpa_set_vq_ready() argument 275 vf->vring[qid].ready = ready; in ifcvf_vdpa_set_vq_ready() 278 static bool ifcvf_vdpa_get_vq_ready(struct vdpa_device *vdpa_dev, u16 qid) in ifcvf_vdpa_get_vq_ready() argument 282 return vf->vring[qid].ready; in ifcvf_vdpa_get_vq_ready() [all …]
|
D | ifcvf_base.c | 275 u16 ifcvf_get_vq_state(struct ifcvf_hw *hw, u16 qid) in ifcvf_get_vq_state() argument 283 q_pair_id = qid / (IFCVF_MAX_QUEUE_PAIRS * 2); in ifcvf_get_vq_state() 284 avail_idx_addr = &ifcvf_lm->vring_lm_cfg[q_pair_id].idx_addr[qid % 2]; in ifcvf_get_vq_state() 290 int ifcvf_set_vq_state(struct ifcvf_hw *hw, u16 qid, u16 num) in ifcvf_set_vq_state() argument 297 q_pair_id = qid / (IFCVF_MAX_QUEUE_PAIRS * 2); in ifcvf_set_vq_state() 298 avail_idx_addr = &ifcvf_lm->vring_lm_cfg[q_pair_id].idx_addr[qid % 2]; in ifcvf_set_vq_state() 299 hw->vring[qid].last_avail_idx = num; in ifcvf_set_vq_state() 387 void ifcvf_notify_queue(struct ifcvf_hw *hw, u16 qid) in ifcvf_notify_queue() argument 389 ifc_iowrite16(qid, hw->vring[qid].notify_addr); in ifcvf_notify_queue()
|
/drivers/net/wireless/mediatek/mt76/ |
D | sdio.c | 21 mt76s_alloc_rx_queue(struct mt76_dev *dev, enum mt76_rxq_id qid) in mt76s_alloc_rx_queue() argument 23 struct mt76_queue *q = &dev->q_rx[qid]; in mt76s_alloc_rx_queue() 111 int qid = q - &dev->q_rx[MT_RXQ_MAIN]; in mt76s_process_rx_queue() local 128 if (qid == MT_RXQ_MAIN) in mt76s_process_rx_queue() 134 static void mt76s_process_tx_queue(struct mt76_dev *dev, enum mt76_txq_id qid) in mt76s_process_tx_queue() argument 136 struct mt76_queue *q = dev->q_tx[qid]; in mt76s_process_tx_queue() 147 if (qid == MT_TXQ_MCU) { in mt76s_process_tx_queue() 162 if (qid == MT_TXQ_MCU) in mt76s_process_tx_queue() 165 mt76_txq_schedule(&dev->phy, qid); in mt76s_process_tx_queue() 168 ieee80211_wake_queue(dev->hw, qid); in mt76s_process_tx_queue() [all …]
|
D | tx.c | 228 __mt76_tx_queue_skb(struct mt76_dev *dev, int qid, struct sk_buff *skb, in __mt76_tx_queue_skb() argument 239 idx = dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, sta); in __mt76_tx_queue_skb() 244 q = dev->q_tx[qid]; in __mt76_tx_queue_skb() 261 int qid = skb_get_queue_mapping(skb); in mt76_tx() local 269 if (WARN_ON(qid >= MT_TXQ_PSD)) { in mt76_tx() 270 qid = MT_TXQ_BE; in mt76_tx() 271 skb_set_queue_mapping(skb, qid); in mt76_tx() 277 qid = MT_TXQ_PSD; in mt76_tx() 278 skb_set_queue_mapping(skb, qid); in mt76_tx() 288 q = dev->q_tx[qid]; in mt76_tx() [all …]
|
D | dma.c | 220 mt76_dma_tx_cleanup(struct mt76_dev *dev, enum mt76_txq_id qid, bool flush) in mt76_dma_tx_cleanup() argument 222 struct mt76_queue *q = dev->q_tx[qid]; in mt76_dma_tx_cleanup() 258 qid < IEEE80211_NUM_ACS && q->queued < q->ndesc - 8; in mt76_dma_tx_cleanup() 266 ieee80211_wake_queue(dev->hw, qid); in mt76_dma_tx_cleanup() 317 mt76_dma_tx_queue_skb_raw(struct mt76_dev *dev, enum mt76_txq_id qid, in mt76_dma_tx_queue_skb_raw() argument 320 struct mt76_queue *q = dev->q_tx[qid]; in mt76_dma_tx_queue_skb_raw() 348 mt76_dma_tx_queue_skb(struct mt76_dev *dev, enum mt76_txq_id qid, in mt76_dma_tx_queue_skb() argument 352 struct mt76_queue *q = dev->q_tx[qid]; in mt76_dma_tx_queue_skb() 406 ret = dev->drv->tx_prepare_skb(dev, txwi, qid, wcid, sta, &tx_info); in mt76_dma_tx_queue_skb() 504 mt76_dma_rx_reset(struct mt76_dev *dev, enum mt76_rxq_id qid) in mt76_dma_rx_reset() argument [all …]
|
/drivers/net/ethernet/intel/i40e/ |
D | i40e_xsk.c | 86 u16 qid) in i40e_xsk_pool_enable() argument 95 if (qid >= vsi->num_queue_pairs) in i40e_xsk_pool_enable() 98 if (qid >= netdev->real_num_rx_queues || in i40e_xsk_pool_enable() 99 qid >= netdev->real_num_tx_queues) in i40e_xsk_pool_enable() 106 set_bit(qid, vsi->af_xdp_zc_qps); in i40e_xsk_pool_enable() 111 err = i40e_queue_pair_disable(vsi, qid); in i40e_xsk_pool_enable() 115 err = i40e_realloc_rx_xdp_bi(vsi->rx_rings[qid], true); in i40e_xsk_pool_enable() 119 err = i40e_queue_pair_enable(vsi, qid); in i40e_xsk_pool_enable() 124 err = i40e_xsk_wakeup(vsi->netdev, qid, XDP_WAKEUP_RX); in i40e_xsk_pool_enable() 140 static int i40e_xsk_pool_disable(struct i40e_vsi *vsi, u16 qid) in i40e_xsk_pool_disable() argument [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | params.h | 44 u16 qid, in mlx5e_qid_get_ch_if_in_group() argument 49 int ch = qid - nch * group; in mlx5e_qid_get_ch_if_in_group() 59 u16 qid, in mlx5e_qid_get_ch_and_group() argument 65 *ix = qid % nch; in mlx5e_qid_get_ch_and_group() 66 *group = qid / nch; in mlx5e_qid_get_ch_and_group() 70 struct mlx5e_params *params, u64 qid) in mlx5e_qid_validate() argument 72 return qid < params->num_channels * profile->rq_groups; in mlx5e_qid_validate()
|
/drivers/net/ethernet/intel/ice/ |
D | ice_xsk.c | 291 static void ice_xsk_remove_pool(struct ice_vsi *vsi, u16 qid) in ice_xsk_remove_pool() argument 293 vsi->xsk_pools[qid] = NULL; in ice_xsk_remove_pool() 310 static int ice_xsk_pool_disable(struct ice_vsi *vsi, u16 qid) in ice_xsk_pool_disable() argument 312 if (!vsi->xsk_pools || qid >= vsi->num_xsk_pools || in ice_xsk_pool_disable() 313 !vsi->xsk_pools[qid]) in ice_xsk_pool_disable() 316 xsk_pool_dma_unmap(vsi->xsk_pools[qid], ICE_RX_DMA_ATTR); in ice_xsk_pool_disable() 317 ice_xsk_remove_pool(vsi, qid); in ice_xsk_pool_disable() 331 ice_xsk_pool_enable(struct ice_vsi *vsi, struct xsk_buff_pool *pool, u16 qid) in ice_xsk_pool_enable() argument 340 if (qid >= vsi->num_xsk_pools) in ice_xsk_pool_enable() 347 if (vsi->xsk_pools && vsi->xsk_pools[qid]) in ice_xsk_pool_enable() [all …]
|
/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_sriov.c | 235 struct bnx2x_virtf *vf, int qid, in bnx2x_vf_queue_create() argument 241 DP(BNX2X_MSG_IOV, "vf[%d:%d]\n", vf->abs_vfid, qid); in bnx2x_vf_queue_create() 245 q_params->q_obj = &bnx2x_vfq(vf, qid, sp_obj); in bnx2x_vf_queue_create() 268 bnx2x_vf_igu_ack_sb(bp, vf, vf_igu_sb(vf, bnx2x_vfq(vf, qid, sb_idx)), in bnx2x_vf_queue_create() 275 int qid) in bnx2x_vf_queue_destroy() argument 287 q_params.q_obj = &bnx2x_vfq(vf, qid, sp_obj); in bnx2x_vf_queue_destroy() 307 if (bnx2x_vfq(vf, qid, cxt)) { in bnx2x_vf_queue_destroy() 308 bnx2x_vfq(vf, qid, cxt)->ustorm_ag_context.cdu_usage = 0; in bnx2x_vf_queue_destroy() 309 bnx2x_vfq(vf, qid, cxt)->xstorm_ag_context.cdu_reserved = 0; in bnx2x_vf_queue_destroy() 335 int qid, bool drv_only, int type) in bnx2x_vf_vlan_mac_clear() argument [all …]
|
/drivers/net/hyperv/ |
D | netvsc_trace.h | 139 __field( u16, qid ) 146 __entry->qid = chan->offermsg.offer.sub_channel_index; 152 __get_str(name), __entry->qid, 164 __field( u16, qid ) 169 __entry->qid = chan->offermsg.offer.sub_channel_index; 173 __get_str(name), __entry->qid,
|