Home
last modified time | relevance | path

Searched refs:rspq (Results 1 – 24 of 24) sorted by relevance

/drivers/net/ethernet/brocade/bna/
Dbfa_msgq.c319 static void bfa_msgq_rspq_dbell(struct bfa_msgq_rspq *rspq);
330 bfa_fsm_state_decl(rspq, stopped, struct bfa_msgq_rspq, enum rspq_event);
331 bfa_fsm_state_decl(rspq, init_wait, struct bfa_msgq_rspq,
333 bfa_fsm_state_decl(rspq, ready, struct bfa_msgq_rspq, enum rspq_event);
334 bfa_fsm_state_decl(rspq, dbell_wait, struct bfa_msgq_rspq,
338 rspq_sm_stopped_entry(struct bfa_msgq_rspq *rspq) in rspq_sm_stopped_entry() argument
340 rspq->producer_index = 0; in rspq_sm_stopped_entry()
341 rspq->consumer_index = 0; in rspq_sm_stopped_entry()
342 rspq->flags = 0; in rspq_sm_stopped_entry()
346 rspq_sm_stopped(struct bfa_msgq_rspq *rspq, enum rspq_event event) in rspq_sm_stopped() argument
[all …]
Dbfa_msgq.h113 struct bfa_msgq_rspq rspq; member
Dbfi.h429 struct bfi_msgq rspq; member
/drivers/net/ethernet/chelsio/cxgb4vf/
Dsge.c1571 struct adapter *adapter = rxq->rspq.adapter; in do_gro()
1576 skb = napi_get_frags(&rxq->rspq.napi); in do_gro()
1588 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro()
1595 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro()
1613 int t4vf_ethrx_handler(struct sge_rspq *rspq, const __be64 *rsp, in t4vf_ethrx_handler() argument
1619 (rspq->netdev->features & NETIF_F_RXCSUM); in t4vf_ethrx_handler()
1620 struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq); in t4vf_ethrx_handler()
1621 struct adapter *adapter = rspq->adapter; in t4vf_ethrx_handler()
1629 (rspq->netdev->features & NETIF_F_GRO) && csum_ok && in t4vf_ethrx_handler()
1645 skb->protocol = eth_type_trans(skb, rspq->netdev); in t4vf_ethrx_handler()
[all …]
Dcxgb4vf_main.c342 &s->ethrxq[rxq].rspq); in request_msix_queue_irqs()
351 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq); in request_msix_queue_irqs()
368 &s->ethrxq[rxq].rspq); in free_msix_queue_irqs()
374 static void qenable(struct sge_rspq *rspq) in qenable() argument
376 napi_enable(&rspq->napi); in qenable()
382 t4_write_reg(rspq->adapter, T4VF_SGE_BASE_ADDR + SGE_VF_GTS, in qenable()
384 SEINTARM_V(rspq->intr_params) | in qenable()
385 INGRESSQID_V(rspq->cntxt_id)); in qenable()
397 qenable(&s->ethrxq[rxq].rspq); in enable_rx()
421 napi_disable(&s->ethrxq[rxq].rspq.napi); in quiesce_rx()
[all …]
Dt4vf_hw.c933 int start, int n, const u16 *rspq, int nrspq) in t4vf_config_rss_range() argument
935 const u16 *rsp = rspq; in t4vf_config_rss_range()
936 const u16 *rsp_end = rspq+nrspq; in t4vf_config_rss_range()
995 rsp = rspq; in t4vf_config_rss_range()
Dadapter.h211 struct sge_rspq rspq; /* Response Queue */ member
/drivers/scsi/bfa/
Dbfa_hw_cb.c53 bfa_hwcb_rspq_ack_msix(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwcb_rspq_ack_msix() argument
55 writel(__HFN_INT_RME_Q0 << RME_Q_NUM(bfa_ioc_pcifn(&bfa->ioc), rspq), in bfa_hwcb_rspq_ack_msix()
58 if (bfa_rspq_ci(bfa, rspq) == ci) in bfa_hwcb_rspq_ack_msix()
61 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwcb_rspq_ack_msix()
62 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwcb_rspq_ack_msix()
67 bfa_hwcb_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwcb_rspq_ack() argument
69 if (bfa_rspq_ci(bfa, rspq) == ci) in bfa_hwcb_rspq_ack()
72 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwcb_rspq_ack()
73 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwcb_rspq_ack()
Dbfa_hw_ct.c74 bfa_hwct_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwct_rspq_ack() argument
78 r32 = readl(bfa->iocfc.bfa_regs.rme_q_ctrl[rspq]); in bfa_hwct_rspq_ack()
79 writel(r32, bfa->iocfc.bfa_regs.rme_q_ctrl[rspq]); in bfa_hwct_rspq_ack()
81 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwct_rspq_ack()
82 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwct_rspq_ack()
92 bfa_hwct2_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwct2_rspq_ack() argument
94 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwct2_rspq_ack()
95 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwct2_rspq_ack()
Dbfa.h187 void (*hw_rspq_ack)(struct bfa_s *bfa, int rspq, u32 ci);
321 void bfa_hwcb_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci);
333 void bfa_hwct_reqq_ack(struct bfa_s *bfa, int rspq);
334 void bfa_hwct_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci);
335 void bfa_hwct2_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci);
Dbfi.h600 struct bfi_msgq_s rspq; member
/drivers/net/ethernet/chelsio/cxgb3/
Dsge.c174 return container_of(q, struct sge_qset, rspq); in rspq_to_qset()
652 memset(&q->rspq, 0, sizeof(q->rspq)); in t3_reset_qset()
707 if (q->rspq.desc) { in t3_free_qset()
709 t3_sge_disable_rspcntxt(adapter, q->rspq.cntxt_id); in t3_free_qset()
712 q->rspq.size * sizeof(struct rsp_desc), in t3_free_qset()
713 q->rspq.desc, q->rspq.phys_addr); in t3_free_qset()
728 qs->rspq.cntxt_id = id; in init_qset_cntxt()
1833 struct sge_rspq *q = &qs->rspq; in ofld_poll()
2258 struct sge_rspq *q = &qs->rspq; in process_responses()
2433 t3_write_reg(adap, A_SG_GTS, V_RSPQ(qs->rspq.cntxt_id) | in napi_rx_handler()
[all …]
Dadapter.h203 struct sge_rspq rspq; member
Dcommon.h699 const u8 * cpus, const u16 *rspq);
748 unsigned int size, int rspq, int ovfl_mode,
Dcxgb3_main.c413 rspq.polling), 0, in request_msix_data_irqs()
449 while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) { in await_mgmt_replies()
462 unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts; in init_tp_parity()
1171 struct sge_rspq *q = &adap->sge.qs[i].rspq; in t3_synchronize_rx()
1277 adap->sge.qs[0].rspq. in cxgb_up()
2622 t3_intr_handler(adapter, qs->rspq.polling) (0, source); in cxgb_netpoll()
Dt3_hw.c2330 unsigned int size, int rspq, int ovfl_mode, in t3_sge_init_cqcntxt() argument
2343 V_CQ_BASE_HI((u32) base_addr) | V_CQ_RSPQ(rspq) | in t3_sge_init_cqcntxt()
2501 const u8 * cpus, const u16 *rspq) in t3_config_rss() argument
2517 if (rspq) in t3_config_rss()
2520 (i << 16) | rspq[q_idx++]); in t3_config_rss()
2521 if (rspq[q_idx] == 0xffff) in t3_config_rss()
Dt3_cpl.h1441 __u8 rspq:3; member
1445 __u8 rspq:3; member
/drivers/staging/slicoss/
Dslicoss.c1187 struct slic_rspqueue *rspq = &adapter->rspqueue; in slic_rspqueue_free() local
1189 for (i = 0; i < rspq->num_pages; i++) { in slic_rspqueue_free()
1190 if (rspq->vaddr[i]) { in slic_rspqueue_free()
1192 rspq->vaddr[i], rspq->paddr[i]); in slic_rspqueue_free()
1194 rspq->vaddr[i] = NULL; in slic_rspqueue_free()
1195 rspq->paddr[i] = 0; in slic_rspqueue_free()
1197 rspq->offset = 0; in slic_rspqueue_free()
1198 rspq->pageindex = 0; in slic_rspqueue_free()
1199 rspq->rspbuf = NULL; in slic_rspqueue_free()
1205 struct slic_rspqueue *rspq = &adapter->rspqueue; in slic_rspqueue_init() local
[all …]
/drivers/net/ethernet/chelsio/cxgb4/
Dcxgb4_debugfs.c2284 rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A"); in sge_qinfo_show()
2299 R("RspQ ID:", rspq.abs_id); in sge_qinfo_show()
2300 R("RspQ size:", rspq.size); in sge_qinfo_show()
2301 R("RspQE size:", rspq.iqe_len); in sge_qinfo_show()
2302 R("RspQ CIDX:", rspq.cidx); in sge_qinfo_show()
2303 R("RspQ Gen:", rspq.gen); in sge_qinfo_show()
2304 S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq)); in sge_qinfo_show()
2306 adap->sge.counter_val[rx[i].rspq.pktcnt_idx]); in sge_qinfo_show()
2342 R("RspQ ID:", rspq.abs_id); in sge_qinfo_show()
2343 R("RspQ size:", rspq.size); in sge_qinfo_show()
[all …]
Dcxgb4_main.c695 struct sge_ofld_rxq *rxq = container_of(q, struct sge_ofld_rxq, rspq); in uldrx_handler()
797 &s->ethrxq[ethqidx].rspq); in request_msix_queue_irqs()
806 &s->ofldrxq[ofldqidx].rspq); in request_msix_queue_irqs()
815 &s->rdmarxq[rdmaqidx].rspq); in request_msix_queue_irqs()
824 &s->rdmaciq[rdmaciqqidx].rspq); in request_msix_queue_irqs()
834 &s->rdmaciq[rdmaciqqidx].rspq); in request_msix_queue_irqs()
837 &s->rdmarxq[rdmaqidx].rspq); in request_msix_queue_irqs()
840 &s->ofldrxq[ofldqidx].rspq); in request_msix_queue_irqs()
843 &s->ethrxq[ethqidx].rspq); in request_msix_queue_irqs()
855 free_irq(adap->msix_info[msi_index++].vec, &s->ethrxq[i].rspq); in free_msix_queue_irqs()
[all …]
Dcxgb4_ethtool.c705 e->rx_mini_pending = s->ethrxq[pi->first_qset].rspq.size; in get_sge_param()
730 s->ethrxq[pi->first_qset + i].rspq.size = e->rx_mini_pending; in set_sge_param()
752 err = cxgb4_set_rspq_intr_params(&q->rspq, us, cnt); in set_rx_intr_params()
767 q->rspq.adaptive_rx = adaptive_rx; in set_adaptive_rx_setting()
778 return q->rspq.adaptive_rx; in get_adaptive_rx_setting()
792 const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq; in get_coalesce()
Dsge.c1848 struct adapter *adapter = rxq->rspq.adap; in do_gro()
1854 skb = napi_get_frags(&rxq->rspq.napi); in do_gro()
1866 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro()
1867 skb_mark_napi_id(skb, &rxq->rspq.napi); in do_gro()
1872 if (rxq->rspq.netdev->features & NETIF_F_RXHASH) in do_gro()
1880 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro()
1903 struct sge_eth_rxq *rxq = container_of(q, struct sge_eth_rxq, rspq); in t4_ethrx_handler()
2059 struct sge_eth_rxq *rxq = container_of(q, struct sge_eth_rxq, rspq); in process_responses()
2355 if (napi_reschedule(&rxq->rspq.napi)) in sge_rx_timer_cb()
2855 if (q->rspq.desc) in t4_free_ofld_rxqs()
[all …]
Dcxgb4.h560 struct sge_rspq rspq; member
573 struct sge_rspq rspq; member
1285 int start, int n, const u16 *rspq, unsigned int nrspq);
Dt4_hw.c4464 int start, int n, const u16 *rspq, unsigned int nrspq) in t4_config_rss_range() argument
4467 const u16 *rsp = rspq; in t4_config_rss_range()
4468 const u16 *rsp_end = rspq + nrspq; in t4_config_rss_range()
4493 rsp = rspq; in t4_config_rss_range()
4496 rsp = rspq; in t4_config_rss_range()
4499 rsp = rspq; in t4_config_rss_range()