/drivers/net/ethernet/brocade/bna/ |
D | bfa_msgq.c | 319 static void bfa_msgq_rspq_dbell(struct bfa_msgq_rspq *rspq); 330 bfa_fsm_state_decl(rspq, stopped, struct bfa_msgq_rspq, enum rspq_event); 331 bfa_fsm_state_decl(rspq, init_wait, struct bfa_msgq_rspq, 333 bfa_fsm_state_decl(rspq, ready, struct bfa_msgq_rspq, enum rspq_event); 334 bfa_fsm_state_decl(rspq, dbell_wait, struct bfa_msgq_rspq, 338 rspq_sm_stopped_entry(struct bfa_msgq_rspq *rspq) in rspq_sm_stopped_entry() argument 340 rspq->producer_index = 0; in rspq_sm_stopped_entry() 341 rspq->consumer_index = 0; in rspq_sm_stopped_entry() 342 rspq->flags = 0; in rspq_sm_stopped_entry() 346 rspq_sm_stopped(struct bfa_msgq_rspq *rspq, enum rspq_event event) in rspq_sm_stopped() argument [all …]
|
D | bfa_msgq.h | 113 struct bfa_msgq_rspq rspq; member
|
D | bfi.h | 429 struct bfi_msgq rspq; member
|
/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 1571 struct adapter *adapter = rxq->rspq.adapter; in do_gro() 1576 skb = napi_get_frags(&rxq->rspq.napi); in do_gro() 1588 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro() 1595 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro() 1613 int t4vf_ethrx_handler(struct sge_rspq *rspq, const __be64 *rsp, in t4vf_ethrx_handler() argument 1619 (rspq->netdev->features & NETIF_F_RXCSUM); in t4vf_ethrx_handler() 1620 struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq); in t4vf_ethrx_handler() 1621 struct adapter *adapter = rspq->adapter; in t4vf_ethrx_handler() 1629 (rspq->netdev->features & NETIF_F_GRO) && csum_ok && in t4vf_ethrx_handler() 1645 skb->protocol = eth_type_trans(skb, rspq->netdev); in t4vf_ethrx_handler() [all …]
|
D | cxgb4vf_main.c | 342 &s->ethrxq[rxq].rspq); in request_msix_queue_irqs() 351 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq); in request_msix_queue_irqs() 368 &s->ethrxq[rxq].rspq); in free_msix_queue_irqs() 374 static void qenable(struct sge_rspq *rspq) in qenable() argument 376 napi_enable(&rspq->napi); in qenable() 382 t4_write_reg(rspq->adapter, T4VF_SGE_BASE_ADDR + SGE_VF_GTS, in qenable() 384 SEINTARM_V(rspq->intr_params) | in qenable() 385 INGRESSQID_V(rspq->cntxt_id)); in qenable() 397 qenable(&s->ethrxq[rxq].rspq); in enable_rx() 421 napi_disable(&s->ethrxq[rxq].rspq.napi); in quiesce_rx() [all …]
|
D | t4vf_hw.c | 933 int start, int n, const u16 *rspq, int nrspq) in t4vf_config_rss_range() argument 935 const u16 *rsp = rspq; in t4vf_config_rss_range() 936 const u16 *rsp_end = rspq+nrspq; in t4vf_config_rss_range() 995 rsp = rspq; in t4vf_config_rss_range()
|
D | adapter.h | 211 struct sge_rspq rspq; /* Response Queue */ member
|
/drivers/scsi/bfa/ |
D | bfa_hw_cb.c | 53 bfa_hwcb_rspq_ack_msix(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwcb_rspq_ack_msix() argument 55 writel(__HFN_INT_RME_Q0 << RME_Q_NUM(bfa_ioc_pcifn(&bfa->ioc), rspq), in bfa_hwcb_rspq_ack_msix() 58 if (bfa_rspq_ci(bfa, rspq) == ci) in bfa_hwcb_rspq_ack_msix() 61 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwcb_rspq_ack_msix() 62 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwcb_rspq_ack_msix() 67 bfa_hwcb_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwcb_rspq_ack() argument 69 if (bfa_rspq_ci(bfa, rspq) == ci) in bfa_hwcb_rspq_ack() 72 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwcb_rspq_ack() 73 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwcb_rspq_ack()
|
D | bfa_hw_ct.c | 74 bfa_hwct_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwct_rspq_ack() argument 78 r32 = readl(bfa->iocfc.bfa_regs.rme_q_ctrl[rspq]); in bfa_hwct_rspq_ack() 79 writel(r32, bfa->iocfc.bfa_regs.rme_q_ctrl[rspq]); in bfa_hwct_rspq_ack() 81 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwct_rspq_ack() 82 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwct_rspq_ack() 92 bfa_hwct2_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwct2_rspq_ack() argument 94 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwct2_rspq_ack() 95 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwct2_rspq_ack()
|
D | bfa.h | 187 void (*hw_rspq_ack)(struct bfa_s *bfa, int rspq, u32 ci); 321 void bfa_hwcb_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci); 333 void bfa_hwct_reqq_ack(struct bfa_s *bfa, int rspq); 334 void bfa_hwct_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci); 335 void bfa_hwct2_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci);
|
D | bfi.h | 600 struct bfi_msgq_s rspq; member
|
/drivers/net/ethernet/chelsio/cxgb3/ |
D | sge.c | 174 return container_of(q, struct sge_qset, rspq); in rspq_to_qset() 652 memset(&q->rspq, 0, sizeof(q->rspq)); in t3_reset_qset() 707 if (q->rspq.desc) { in t3_free_qset() 709 t3_sge_disable_rspcntxt(adapter, q->rspq.cntxt_id); in t3_free_qset() 712 q->rspq.size * sizeof(struct rsp_desc), in t3_free_qset() 713 q->rspq.desc, q->rspq.phys_addr); in t3_free_qset() 728 qs->rspq.cntxt_id = id; in init_qset_cntxt() 1833 struct sge_rspq *q = &qs->rspq; in ofld_poll() 2258 struct sge_rspq *q = &qs->rspq; in process_responses() 2433 t3_write_reg(adap, A_SG_GTS, V_RSPQ(qs->rspq.cntxt_id) | in napi_rx_handler() [all …]
|
D | adapter.h | 203 struct sge_rspq rspq; member
|
D | common.h | 699 const u8 * cpus, const u16 *rspq); 748 unsigned int size, int rspq, int ovfl_mode,
|
D | cxgb3_main.c | 413 rspq.polling), 0, in request_msix_data_irqs() 449 while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) { in await_mgmt_replies() 462 unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts; in init_tp_parity() 1171 struct sge_rspq *q = &adap->sge.qs[i].rspq; in t3_synchronize_rx() 1277 adap->sge.qs[0].rspq. in cxgb_up() 2622 t3_intr_handler(adapter, qs->rspq.polling) (0, source); in cxgb_netpoll()
|
D | t3_hw.c | 2330 unsigned int size, int rspq, int ovfl_mode, in t3_sge_init_cqcntxt() argument 2343 V_CQ_BASE_HI((u32) base_addr) | V_CQ_RSPQ(rspq) | in t3_sge_init_cqcntxt() 2501 const u8 * cpus, const u16 *rspq) in t3_config_rss() argument 2517 if (rspq) in t3_config_rss() 2520 (i << 16) | rspq[q_idx++]); in t3_config_rss() 2521 if (rspq[q_idx] == 0xffff) in t3_config_rss()
|
D | t3_cpl.h | 1441 __u8 rspq:3; member 1445 __u8 rspq:3; member
|
/drivers/staging/slicoss/ |
D | slicoss.c | 1187 struct slic_rspqueue *rspq = &adapter->rspqueue; in slic_rspqueue_free() local 1189 for (i = 0; i < rspq->num_pages; i++) { in slic_rspqueue_free() 1190 if (rspq->vaddr[i]) { in slic_rspqueue_free() 1192 rspq->vaddr[i], rspq->paddr[i]); in slic_rspqueue_free() 1194 rspq->vaddr[i] = NULL; in slic_rspqueue_free() 1195 rspq->paddr[i] = 0; in slic_rspqueue_free() 1197 rspq->offset = 0; in slic_rspqueue_free() 1198 rspq->pageindex = 0; in slic_rspqueue_free() 1199 rspq->rspbuf = NULL; in slic_rspqueue_free() 1205 struct slic_rspqueue *rspq = &adapter->rspqueue; in slic_rspqueue_init() local [all …]
|
/drivers/net/ethernet/chelsio/cxgb4/ |
D | cxgb4_debugfs.c | 2284 rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A"); in sge_qinfo_show() 2299 R("RspQ ID:", rspq.abs_id); in sge_qinfo_show() 2300 R("RspQ size:", rspq.size); in sge_qinfo_show() 2301 R("RspQE size:", rspq.iqe_len); in sge_qinfo_show() 2302 R("RspQ CIDX:", rspq.cidx); in sge_qinfo_show() 2303 R("RspQ Gen:", rspq.gen); in sge_qinfo_show() 2304 S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq)); in sge_qinfo_show() 2306 adap->sge.counter_val[rx[i].rspq.pktcnt_idx]); in sge_qinfo_show() 2342 R("RspQ ID:", rspq.abs_id); in sge_qinfo_show() 2343 R("RspQ size:", rspq.size); in sge_qinfo_show() [all …]
|
D | cxgb4_main.c | 695 struct sge_ofld_rxq *rxq = container_of(q, struct sge_ofld_rxq, rspq); in uldrx_handler() 797 &s->ethrxq[ethqidx].rspq); in request_msix_queue_irqs() 806 &s->ofldrxq[ofldqidx].rspq); in request_msix_queue_irqs() 815 &s->rdmarxq[rdmaqidx].rspq); in request_msix_queue_irqs() 824 &s->rdmaciq[rdmaciqqidx].rspq); in request_msix_queue_irqs() 834 &s->rdmaciq[rdmaciqqidx].rspq); in request_msix_queue_irqs() 837 &s->rdmarxq[rdmaqidx].rspq); in request_msix_queue_irqs() 840 &s->ofldrxq[ofldqidx].rspq); in request_msix_queue_irqs() 843 &s->ethrxq[ethqidx].rspq); in request_msix_queue_irqs() 855 free_irq(adap->msix_info[msi_index++].vec, &s->ethrxq[i].rspq); in free_msix_queue_irqs() [all …]
|
D | cxgb4_ethtool.c | 705 e->rx_mini_pending = s->ethrxq[pi->first_qset].rspq.size; in get_sge_param() 730 s->ethrxq[pi->first_qset + i].rspq.size = e->rx_mini_pending; in set_sge_param() 752 err = cxgb4_set_rspq_intr_params(&q->rspq, us, cnt); in set_rx_intr_params() 767 q->rspq.adaptive_rx = adaptive_rx; in set_adaptive_rx_setting() 778 return q->rspq.adaptive_rx; in get_adaptive_rx_setting() 792 const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq; in get_coalesce()
|
D | sge.c | 1848 struct adapter *adapter = rxq->rspq.adap; in do_gro() 1854 skb = napi_get_frags(&rxq->rspq.napi); in do_gro() 1866 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro() 1867 skb_mark_napi_id(skb, &rxq->rspq.napi); in do_gro() 1872 if (rxq->rspq.netdev->features & NETIF_F_RXHASH) in do_gro() 1880 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro() 1903 struct sge_eth_rxq *rxq = container_of(q, struct sge_eth_rxq, rspq); in t4_ethrx_handler() 2059 struct sge_eth_rxq *rxq = container_of(q, struct sge_eth_rxq, rspq); in process_responses() 2355 if (napi_reschedule(&rxq->rspq.napi)) in sge_rx_timer_cb() 2855 if (q->rspq.desc) in t4_free_ofld_rxqs() [all …]
|
D | cxgb4.h | 560 struct sge_rspq rspq; member 573 struct sge_rspq rspq; member 1285 int start, int n, const u16 *rspq, unsigned int nrspq);
|
D | t4_hw.c | 4464 int start, int n, const u16 *rspq, unsigned int nrspq) in t4_config_rss_range() argument 4467 const u16 *rsp = rspq; in t4_config_rss_range() 4468 const u16 *rsp_end = rspq + nrspq; in t4_config_rss_range() 4493 rsp = rspq; in t4_config_rss_range() 4496 rsp = rspq; in t4_config_rss_range() 4499 rsp = rspq; in t4_config_rss_range()
|