/drivers/soc/fsl/qe/ |
D | qe.c | 53 struct device_node *qe; in qe_get_device_node() local 59 qe = of_find_compatible_node(NULL, NULL, "fsl,qe"); in qe_get_device_node() 60 if (qe) in qe_get_device_node() 61 return qe; in qe_get_device_node() 67 struct device_node *qe; in get_qe_base() local 74 qe = qe_get_device_node(); in get_qe_base() 75 if (!qe) in get_qe_base() 78 ret = of_address_to_resource(qe, 0, &res); in get_qe_base() 81 of_node_put(qe); in get_qe_base() 161 struct device_node *qe; in qe_get_brg_clk() local [all …]
|
D | Makefile | 5 obj-$(CONFIG_QUICC_ENGINE)+= qe.o qe_common.o qe_ic.o qe_io.o
|
/drivers/net/ethernet/chelsio/cxgb4/ |
D | sched.c | 82 struct sched_queue_entry *qe; in t4_sched_bind_unbind_op() local 84 qe = (struct sched_queue_entry *)arg; in t4_sched_bind_unbind_op() 92 fw_class = bind ? qe->param.class : FW_SCHED_CLS_NONE; in t4_sched_bind_unbind_op() 93 fw_param = (fw_mnem | FW_PARAMS_PARAM_YZ_V(qe->cntxt_id)); in t4_sched_bind_unbind_op() 137 struct sched_queue_entry *qe; in t4_sched_entry_lookup() local 139 list_for_each_entry(qe, &e->entry_list, list) { in t4_sched_entry_lookup() 140 if (qe->cntxt_id == val) { in t4_sched_entry_lookup() 141 found = qe; in t4_sched_entry_lookup() 173 struct sched_queue_entry *qe = NULL; in cxgb4_sched_queue_lookup() local 181 qe = t4_sched_entry_lookup(pi, SCHED_QUEUE, txq->q.cntxt_id); in cxgb4_sched_queue_lookup() [all …]
|
D | cxgb4_tc_matchall.c | 18 struct ch_sched_queue qe; in cxgb4_matchall_egress_validate() local 66 memset(&qe, 0, sizeof(qe)); in cxgb4_matchall_egress_validate() 67 qe.queue = i; in cxgb4_matchall_egress_validate() 69 e = cxgb4_sched_queue_lookup(dev, &qe); in cxgb4_matchall_egress_validate() 83 struct ch_sched_queue qe; in cxgb4_matchall_tc_bind_queues() local 88 qe.queue = i; in cxgb4_matchall_tc_bind_queues() 89 qe.class = tc; in cxgb4_matchall_tc_bind_queues() 90 ret = cxgb4_sched_class_bind(dev, &qe, SCHED_QUEUE); in cxgb4_matchall_tc_bind_queues() 99 qe.queue = i; in cxgb4_matchall_tc_bind_queues() 100 qe.class = SCHED_CLS_NONE; in cxgb4_matchall_tc_bind_queues() [all …]
|
/drivers/atm/ |
D | firestream.c | 583 static void submit_qentry (struct fs_dev *dev, struct queue *q, struct FS_QENTRY *qe) in submit_qentry() argument 599 if (qe != cqe) { in submit_qentry() 600 fs_dprintk (FS_DEBUG_TXQ, "q mismatch! %p %p\n", qe, cqe); in submit_qentry() 630 struct FS_QENTRY *qe; in submit_queue() local 632 qe = get_qentry (dev, q); in submit_queue() 633 qe->cmd = cmd; in submit_queue() 634 qe->p0 = p1; in submit_queue() 635 qe->p1 = p2; in submit_queue() 636 qe->p2 = p3; in submit_queue() 637 submit_qentry (dev, q, qe); in submit_queue() [all …]
|
D | firestream.h | 50 #define STATUS_CODE(qe) ((qe->cmd >> 22) & 0x3f) argument
|
/drivers/net/ethernet/sun/ |
D | sunqe.c | 837 struct sunqe *qe; in qec_ether_init() local 849 qe = netdev_priv(dev); in qec_ether_init() 856 qe->channel = i; in qec_ether_init() 857 spin_lock_init(&qe->lock); in qec_ether_init() 863 qecp->qes[qe->channel] = qe; in qec_ether_init() 864 qe->dev = dev; in qec_ether_init() 865 qe->parent = qecp; in qec_ether_init() 866 qe->op = op; in qec_ether_init() 869 qe->qcregs = of_ioremap(&op->resource[0], 0, in qec_ether_init() 871 if (!qe->qcregs) { in qec_ether_init() [all …]
|
/drivers/net/ethernet/brocade/bna/ |
D | bna_tx_rx.c | 352 list_for_each_entry(mac, &rxf->mcast_active_q, qe) in bna_rxf_mcmac_get() 356 list_for_each_entry(mac, &rxf->mcast_pending_del_q, qe) in bna_rxf_mcmac_get() 368 list_for_each_entry(mchandle, &rxf->mcast_handle_q, qe) in bna_rxf_mchandle_get() 387 list_add_tail(&mchandle->qe, &rxf->mcast_handle_q); in bna_rxf_mchandle_attach() 410 list_del(&mchandle->qe); in bna_rxf_mcast_del() 427 struct bna_mac, qe); in bna_rxf_mcast_cfg_apply() 429 list_move_tail(&mac->qe, bna_mcam_mod_del_q(rxf->rx->bna)); in bna_rxf_mcast_cfg_apply() 437 struct bna_mac, qe); in bna_rxf_mcast_cfg_apply() 438 list_move_tail(&mac->qe, &rxf->mcast_active_q); in bna_rxf_mcast_cfg_apply() 475 struct bna_mac, qe); in bna_rxf_mcast_cfg_reset() [all …]
|
D | bna_types.h | 256 struct list_head qe; member 440 struct list_head qe; member 459 struct list_head qe; member 563 struct list_head qe; member 687 struct list_head qe; member 774 struct list_head qe; member 866 struct list_head qe; member
|
D | bfa_msgq.c | 62 struct bfa_msgq_cmd_entry, qe); in cmdq_sm_stopped_entry() 63 list_del(&cmdq_ent->qe); in cmdq_sm_stopped_entry() 239 struct bfa_msgq_cmd_entry, qe); in bfa_msgq_cmdq_ci_update() 242 list_del(&cmd->qe); in bfa_msgq_cmdq_ci_update() 632 list_add_tail(&cmd->qe, &msgq->cmdq.pending_q); in bfa_msgq_cmd_post()
|
D | bna_enet.c | 1788 list_add_tail(&ucam_mod->ucmac[i].qe, &ucam_mod->free_q); in bna_ucam_mod_init() 1793 list_add_tail(&ucam_mod->ucmac[i].qe, &ucam_mod->del_q); in bna_ucam_mod_init() 1815 list_add_tail(&mcam_mod->mcmac[i].qe, &mcam_mod->free_q); in bna_mcam_mod_init() 1822 list_add_tail(&mcam_mod->mchandle[i].qe, in bna_mcam_mod_init() 1828 list_add_tail(&mcam_mod->mcmac[i].qe, &mcam_mod->del_q); in bna_mcam_mod_init() 2054 mac = list_first_entry_or_null(head, struct bna_mac, qe); in bna_cam_mod_mac_get() 2056 list_del(&mac->qe); in bna_cam_mod_mac_get() 2067 struct bna_mcam_handle, qe); in bna_mcam_mod_handle_get() 2069 list_del(&handle->qe); in bna_mcam_mod_handle_get() 2078 list_add_tail(&handle->qe, &mcam_mod->free_handle_q); in bna_mcam_mod_handle_put()
|
D | bna.h | 207 list_for_each_entry(__tx, &__tx_mod->tx_active_q, qe) { \ 220 list_for_each_entry(__rx, &__rx_mod->rx_active_q, qe) { \ 242 list_for_each_entry(mac, q, qe) in bna_mac_find()
|
/drivers/scsi/arm/ |
D | msgqueue.c | 55 msgq->qe = NULL; in msgqueue_initialise() 82 struct msgqueue_entry *mq = msgq->qe; in msgqueue_msglength() 85 for (mq = msgq->qe; mq; mq = mq->next) in msgqueue_msglength() 102 for (mq = msgq->qe; mq && msgno; mq = mq->next, msgno--); in msgqueue_getmsg() 133 mqp = &msgq->qe; in msgqueue_addmsg() 152 for (mq = msgq->qe; mq; mq = mqnext) { in msgqueue_flush() 156 msgq->qe = NULL; in msgqueue_flush()
|
D | msgqueue.h | 26 struct msgqueue_entry *qe; member
|
/drivers/scsi/bfa/ |
D | bfa_fcpim.c | 25 list_add_tail(&(__itnim)->qe, &(__itnim)->fcpim->itnim_q) 29 list_del(&(__itnim)->qe); \ 86 list_del(&(__ioim)->qe); \ 87 list_add_tail(&(__ioim)->qe, &(__ioim)->fcpim->ioim_comp_q); \ 338 struct list_head *qe, *qen; in bfa_fcpim_iocdisable() local 343 list_for_each_safe(qe, qen, &fcpim->itnim_q) { in bfa_fcpim_iocdisable() 344 itnim = (struct bfa_itnim_s *) qe; in bfa_fcpim_iocdisable() 425 struct list_head *qe, *qen; in bfa_fcpim_port_iostats() local 430 list_for_each_safe(qe, qen, &fcpim->itnim_q) { in bfa_fcpim_port_iostats() 431 itnim = (struct bfa_itnim_s *) qe; in bfa_fcpim_port_iostats() [all …]
|
D | bfa_fcs.c | 950 struct list_head *qe, *qen; in bfa_fcs_fabric_notify_online() local 962 list_for_each_safe(qe, qen, &fabric->vport_q) { in bfa_fcs_fabric_notify_online() 963 vport = (struct bfa_fcs_vport_s *) qe; in bfa_fcs_fabric_notify_online() 972 struct list_head *qe, *qen; in bfa_fcs_fabric_notify_offline() local 980 list_for_each_safe(qe, qen, &fabric->vport_q) { in bfa_fcs_fabric_notify_offline() 981 vport = (struct bfa_fcs_vport_s *) qe; in bfa_fcs_fabric_notify_offline() 1006 struct list_head *qe, *qen; in bfa_fcs_fabric_stop() local 1010 list_for_each_safe(qe, qen, &fabric->vport_q) { in bfa_fcs_fabric_stop() 1011 vport = (struct bfa_fcs_vport_s *) qe; in bfa_fcs_fabric_stop() 1028 struct list_head *qe, *qen; in bfa_fcs_fabric_delete() local [all …]
|
D | bfa_svc.c | 435 list_add_tail(&fcxp->qe, &mod->fcxp_req_free_q); in claim_fcxps_mem() 438 list_add_tail(&fcxp->qe, &mod->fcxp_rsp_free_q); in claim_fcxps_mem() 516 struct list_head *qe, *qen; in bfa_fcxp_iocdisable() local 522 list_for_each_safe(qe, qen, &mod->fcxp_active_q) { in bfa_fcxp_iocdisable() 523 fcxp = (struct bfa_fcxp_s *) qe; in bfa_fcxp_iocdisable() 547 list_add_tail(&fcxp->qe, &fm->fcxp_active_q); in bfa_fcxp_get() 641 list_del(&fcxp->qe); in bfa_fcxp_put() 644 list_add_tail(&fcxp->qe, &mod->fcxp_req_free_q); in bfa_fcxp_put() 646 list_add_tail(&fcxp->qe, &mod->fcxp_rsp_free_q); in bfa_fcxp_put() 1089 list_add_tail(&wqe->qe, &mod->req_wait_q); in bfa_fcxp_req_rsp_alloc_wait() [all …]
|
D | bfa_fcs_lport.c | 190 struct list_head *qe, *qen; in bfa_fcs_lport_sm_online() local 213 list_for_each_safe(qe, qen, &port->rport_q) { in bfa_fcs_lport_sm_online() 214 rport = (struct bfa_fcs_rport_s *) qe; in bfa_fcs_lport_sm_online() 229 list_for_each_safe(qe, qen, &port->rport_q) { in bfa_fcs_lport_sm_online() 230 rport = (struct bfa_fcs_rport_s *) qe; in bfa_fcs_lport_sm_online() 250 struct list_head *qe, *qen; in bfa_fcs_lport_sm_offline() local 271 list_for_each_safe(qe, qen, &port->rport_q) { in bfa_fcs_lport_sm_offline() 272 rport = (struct bfa_fcs_rport_s *) qe; in bfa_fcs_lport_sm_offline() 284 list_for_each_safe(qe, qen, &port->rport_q) { in bfa_fcs_lport_sm_offline() 285 rport = (struct bfa_fcs_rport_s *) qe; in bfa_fcs_lport_sm_offline() [all …]
|
D | bfa.h | 109 list_add_tail(&(__wqe)->qe, waitq); \ 112 #define bfa_reqq_wcancel(__wqe) list_del(&(__wqe)->qe) 118 list_add_tail(&(__hcb_qe)->qe, &(__bfa)->comp_q); \ 121 #define bfa_cb_dequeue(__hcb_qe) list_del(&(__hcb_qe)->qe) 127 list_add_tail(&(__hcb_qe)->qe, &(__bfa)->comp_q); \ 134 list_add_tail(&(__hcb_qe)->qe, &(__bfa)->comp_q); \ 433 bfa_q_qe_init(&((__qe)->hcb_qe.qe)); \
|
D | bfa_ioc.h | 32 struct list_head qe; member 84 struct list_head qe; /* Queue of DMA elements */ member 94 struct list_head qe; /* Queue of KVA elements */ member 113 list_add_tail(&dm_ptr->qe, &meminfo->dma_info.qe); in bfa_mem_dma_setup() 122 list_add_tail(&kva_ptr->qe, &meminfo->kva_info.qe); in bfa_mem_kva_setup() 229 struct list_head qe; member 272 struct list_head qe; member 359 struct list_head qe; member 370 struct list_head qe; member
|
D | bfa_svc.h | 31 struct list_head qe; /* queue sg page element */ member 40 struct list_head qe; /* queue sg page element */ member 156 struct list_head qe; /* fcxp queue element */ member 200 struct list_head qe; member 291 struct list_head qe; /* queue element */ member 319 struct list_head qe; /* queue element */ member 384 struct list_head qe; /* queue element */ member
|
D | bfa_fcpim.h | 26 struct list_head qe; /* queue element */ member 161 struct list_head qe; /* queue elememt */ member 193 struct list_head qe; member 215 struct list_head qe; /* queue element */ member
|
D | bfa_core.c | 694 struct list_head *waitq, *qe, *qen; in bfa_reqq_resume() local 698 list_for_each_safe(qe, qen, waitq) { in bfa_reqq_resume() 705 list_del(qe); in bfa_reqq_resume() 706 wqe = (struct bfa_reqq_wait_s *) qe; in bfa_reqq_resume() 1774 INIT_LIST_HEAD(&meminfo->dma_info.qe); in bfa_cfg_get_meminfo() 1775 INIT_LIST_HEAD(&meminfo->kva_info.qe); in bfa_cfg_get_meminfo() 1847 list_for_each(dm_qe, &dma_info->qe) { in bfa_attach() 1853 list_for_each(km_qe, &kva_info->qe) { in bfa_attach() 1907 struct list_head *qe; in bfa_comp_process() local 1912 list_for_each_safe(qe, qen, comp_q) { in bfa_comp_process() [all …]
|
/drivers/soc/fsl/ |
D | Makefile | 7 obj-$(CONFIG_QUICC_ENGINE) += qe/ 8 obj-$(CONFIG_CPM) += qe/
|
/drivers/nvme/host/ |
D | rdma.c | 176 static void nvme_rdma_free_qe(struct ib_device *ibdev, struct nvme_rdma_qe *qe, in nvme_rdma_free_qe() argument 179 ib_dma_unmap_single(ibdev, qe->dma, capsule_size, dir); in nvme_rdma_free_qe() 180 kfree(qe->data); in nvme_rdma_free_qe() 183 static int nvme_rdma_alloc_qe(struct ib_device *ibdev, struct nvme_rdma_qe *qe, in nvme_rdma_alloc_qe() argument 186 qe->data = kzalloc(capsule_size, GFP_KERNEL); in nvme_rdma_alloc_qe() 187 if (!qe->data) in nvme_rdma_alloc_qe() 190 qe->dma = ib_dma_map_single(ibdev, qe->data, capsule_size, dir); in nvme_rdma_alloc_qe() 191 if (ib_dma_mapping_error(ibdev, qe->dma)) { in nvme_rdma_alloc_qe() 192 kfree(qe->data); in nvme_rdma_alloc_qe() 193 qe->data = NULL; in nvme_rdma_alloc_qe() [all …]
|