Home
last modified time | relevance | path

Searched refs:rxq (Results 1 – 25 of 26) sorted by relevance

12

/drivers/net/wireless/iwlwifi/
Diwl-rx.c191 struct iwl_rx_queue *rxq = &priv->rxq; in iwl_rx_queue_restock() local
198 spin_lock_irqsave(&rxq->lock, flags); in iwl_rx_queue_restock()
199 write = rxq->write & ~0x7; in iwl_rx_queue_restock()
200 while ((iwl_rx_queue_space(rxq) > 0) && (rxq->free_count)) { in iwl_rx_queue_restock()
202 element = rxq->rx_free.next; in iwl_rx_queue_restock()
207 rxq->bd[rxq->write] = iwl_dma_addr2rbd_ptr(priv, rxb->aligned_dma_addr); in iwl_rx_queue_restock()
208 rxq->queue[rxq->write] = rxb; in iwl_rx_queue_restock()
209 rxq->write = (rxq->write + 1) & RX_QUEUE_MASK; in iwl_rx_queue_restock()
210 rxq->free_count--; in iwl_rx_queue_restock()
212 spin_unlock_irqrestore(&rxq->lock, flags); in iwl_rx_queue_restock()
[all …]
Diwl3945-base.c3521 struct iwl3945_rx_queue *rxq = &priv->rxq; in iwl3945_rx_queue_restock() local
3527 spin_lock_irqsave(&rxq->lock, flags); in iwl3945_rx_queue_restock()
3528 write = rxq->write & ~0x7; in iwl3945_rx_queue_restock()
3529 while ((iwl3945_rx_queue_space(rxq) > 0) && (rxq->free_count)) { in iwl3945_rx_queue_restock()
3531 element = rxq->rx_free.next; in iwl3945_rx_queue_restock()
3536 rxq->bd[rxq->write] = iwl3945_dma_addr2rbd_ptr(priv, rxb->dma_addr); in iwl3945_rx_queue_restock()
3537 rxq->queue[rxq->write] = rxb; in iwl3945_rx_queue_restock()
3538 rxq->write = (rxq->write + 1) & RX_QUEUE_MASK; in iwl3945_rx_queue_restock()
3539 rxq->free_count--; in iwl3945_rx_queue_restock()
3541 spin_unlock_irqrestore(&rxq->lock, flags); in iwl3945_rx_queue_restock()
[all …]
Diwl-core.h230 void iwl_rx_queue_free(struct iwl_priv *priv, struct iwl_rx_queue *rxq);
235 void iwl_rx_queue_reset(struct iwl_priv *priv, struct iwl_rx_queue *rxq);
237 int iwl_rx_init(struct iwl_priv *priv, struct iwl_rx_queue *rxq);
Diwl-core.c196 struct iwl_rx_queue *rxq = &priv->rxq; in iwl_hw_nic_init() local
210 if (!rxq->bd) { in iwl_hw_nic_init()
217 iwl_rx_queue_reset(priv, rxq); in iwl_hw_nic_init()
221 iwl_rx_init(priv, rxq); in iwl_hw_nic_init()
225 rxq->need_update = 1; in iwl_hw_nic_init()
226 iwl_rx_queue_update_write_ptr(priv, rxq); in iwl_hw_nic_init()
Diwl-agn.c1124 struct iwl_rx_queue *rxq = &priv->rxq; in iwl_rx_handle() local
1133 r = le16_to_cpu(rxq->rb_stts->closed_rb_num) & 0x0FFF; in iwl_rx_handle()
1134 i = rxq->read; in iwl_rx_handle()
1140 if (iwl_rx_queue_space(rxq) > (RX_QUEUE_SIZE / 2)) in iwl_rx_handle()
1144 rxb = rxq->queue[i]; in iwl_rx_handle()
1151 rxq->queue[i] = NULL; in iwl_rx_handle()
1211 spin_lock_irqsave(&rxq->lock, flags); in iwl_rx_handle()
1212 list_add_tail(&rxb->list, &priv->rxq.rx_used); in iwl_rx_handle()
1213 spin_unlock_irqrestore(&rxq->lock, flags); in iwl_rx_handle()
1220 priv->rxq.read = i; in iwl_rx_handle()
[all …]
Diwl-3945.c975 static int iwl3945_rx_init(struct iwl3945_priv *priv, struct iwl3945_rx_queue *rxq) in iwl3945_rx_init() argument
987 iwl3945_write_direct32(priv, FH_RCSR_RBD_BASE(0), rxq->dma_addr); in iwl3945_rx_init()
1096 struct iwl3945_rx_queue *rxq = &priv->rxq; in iwl3945_hw_nic_init() local
1188 if (!rxq->bd) { in iwl3945_hw_nic_init()
1195 iwl3945_rx_queue_reset(priv, rxq); in iwl3945_hw_nic_init()
1199 iwl3945_rx_init(priv, rxq); in iwl3945_hw_nic_init()
1213 iwl3945_write_direct32(priv, FH_RCSR_WPTR(0), rxq->write & ~7); in iwl3945_hw_nic_init()
Diwl-3945.h575 struct iwl3945_rx_queue *rxq);
817 struct iwl3945_rx_queue rxq; member
Diwl-dev.h868 struct iwl_rx_queue rxq; member
/drivers/atm/
Dambassador.c687 amb_rxq * rxq = &dev->rxq[pool]; in rx_give() local
692 spin_lock_irqsave (&rxq->lock, flags); in rx_give()
694 if (rxq->pending < rxq->maximum) { in rx_give()
695 PRINTD (DBG_RX, "RX in slot %p", rxq->in.ptr); in rx_give()
697 *rxq->in.ptr = *rx; in rx_give()
698 rxq->pending++; in rx_give()
699 rxq->in.ptr = NEXTQ (rxq->in.ptr, rxq->in.start, rxq->in.limit); in rx_give()
701 wr_mem (dev, offsetof(amb_mem, mb.adapter.rx_address[pool]), virt_to_bus (rxq->in.ptr)); in rx_give()
703 spin_unlock_irqrestore (&rxq->lock, flags); in rx_give()
706 spin_unlock_irqrestore (&rxq->lock, flags); in rx_give()
[all …]
Dfore200e.c1179 struct host_rxq* rxq = &fore200e->host_rxq; in fore200e_rx_irq() local
1186 entry = &rxq->host_entry[ rxq->head ]; in fore200e_rx_irq()
1217 FORE200E_NEXT_ENTRY(rxq->head, QUEUE_SIZE_RX); in fore200e_rx_irq()
2215 struct host_rxq* rxq = &fore200e->host_rxq; in fore200e_init_rx_queue() local
2223 &rxq->status, in fore200e_init_rx_queue()
2232 &rxq->rpd, in fore200e_init_rx_queue()
2237 fore200e->bus->dma_chunk_free(fore200e, &rxq->status); in fore200e_init_rx_queue()
2247 rxq->host_entry[ i ].status = in fore200e_init_rx_queue()
2248 FORE200E_INDEX(rxq->status.align_addr, enum status, i); in fore200e_init_rx_queue()
2249 rxq->host_entry[ i ].rpd = in fore200e_init_rx_queue()
[all …]
Dambassador.h639 amb_rxq rxq[NUM_RX_POOLS]; member
/drivers/net/
Dmv643xx_eth.c393 struct rx_queue rxq[8]; member
429 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) in rxq_to_mp() argument
431 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp()
439 static void rxq_enable(struct rx_queue *rxq) in rxq_enable() argument
441 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_enable()
442 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); in rxq_enable()
445 static void rxq_disable(struct rx_queue *rxq) in rxq_disable() argument
447 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_disable()
448 u8 mask = 1 << rxq->index; in rxq_disable()
496 static int rxq_process(struct rx_queue *rxq, int budget) in rxq_process() argument
[all …]
Dibmveth.h49 #define h_register_logical_lan(ua, buflst, rxq, fltlst, mac) \ argument
50 plpar_hcall_norets(H_REGISTER_LOGICAL_LAN, ua, buflst, rxq, fltlst, mac)
Dsky2.c1132 unsigned rxq = rxqaddr[sky2->port]; in sky2_rx_stop() local
1136 sky2_write8(hw, RB_ADDR(rxq, RB_CTRL), RB_DIS_OP_MD); in sky2_rx_stop()
1139 if (sky2_read8(hw, RB_ADDR(rxq, Q_RSL)) in sky2_rx_stop()
1140 == sky2_read8(hw, RB_ADDR(rxq, Q_RL))) in sky2_rx_stop()
1146 sky2_write32(hw, Q_ADDR(rxq, Q_CSR), BMU_RST_SET | BMU_FIFO_RST); in sky2_rx_stop()
1149 sky2_write32(hw, Y2_QADDR(rxq, PREF_UNIT_CTRL), PREF_UNIT_RST_SET); in sky2_rx_stop()
1289 static inline void sky2_rx_update(struct sky2_port *sky2, unsigned rxq) in sky2_rx_update() argument
1291 sky2_put_idx(sky2->hw, rxq, sky2->rx_put); in sky2_rx_update()
1307 unsigned rxq = rxqaddr[sky2->port]; in sky2_rx_start() local
1311 sky2_qset(hw, rxq); in sky2_rx_start()
[all …]
Dxen-netfront.c811 struct sk_buff_head *rxq) in handle_incoming_queue() argument
816 while ((skb = __skb_dequeue(rxq)) != NULL) { in handle_incoming_queue()
859 struct sk_buff_head rxq; in xennet_poll() local
868 skb_queue_head_init(&rxq); in xennet_poll()
956 __skb_queue_tail(&rxq, skb); in xennet_poll()
964 work_done -= handle_incoming_queue(dev, &rxq); in xennet_poll()
/drivers/net/usb/
Dusbnet.c323 spin_lock_irqsave (&dev->rxq.lock, lockflags); in rx_submit()
346 __skb_queue_tail (&dev->rxq, skb); in rx_submit()
353 spin_unlock_irqrestore (&dev->rxq.lock, lockflags); in rx_submit()
455 defer_bh(dev, skb, &dev->rxq); in rx_complete()
541 (void) unlink_urbs (dev, &dev->rxq); in usbnet_unlink_rx_urbs()
569 temp = unlink_urbs (dev, &dev->txq) + unlink_urbs (dev, &dev->rxq); in usbnet_stop()
572 while (!skb_queue_empty(&dev->rxq) in usbnet_stop()
812 unlink_urbs (dev, &dev->rxq); in kevent()
1037 if ((dev->txq.qlen + dev->rxq.qlen + dev->done.qlen) == 0) { in usbnet_bh()
1046 int temp = dev->rxq.qlen; in usbnet_bh()
[all …]
/drivers/net/benet/
Dbe_main.c602 struct be_queue_info *rxq = &adapter->rx_obj.q; in get_rx_page_info() local
611 atomic_dec(&rxq->used); in get_rx_page_info()
619 struct be_queue_info *rxq = &adapter->rx_obj.q; in be_rx_compl_discard() local
630 index_inc(&rxq_idx, rxq->len); in be_rx_compl_discard()
641 struct be_queue_info *rxq = &adapter->rx_obj.q; in skb_fill_rx_data() local
687 index_inc(&rxq_idx, rxq->len); in skb_fill_rx_data()
762 struct be_queue_info *rxq = &adapter->rx_obj.q; in be_rx_compl_process_lro() local
782 index_inc(&rxq_idx, rxq->len); in be_rx_compl_process_lro()
838 struct be_queue_info *rxq = &adapter->rx_obj.q; in be_post_rx_frags() local
845 page_info = &page_info_tbl[rxq->head]; in be_post_rx_frags()
[all …]
Dbe_cmds.c455 struct be_queue_info *rxq, u16 cq_id, u16 frag_size, in be_cmd_rxq_create() argument
460 struct be_dma_mem *q_mem = &rxq->dma_mem; in be_cmd_rxq_create()
482 rxq->id = le16_to_cpu(resp->id); in be_cmd_rxq_create()
483 rxq->created = true; in be_cmd_rxq_create()
Dbe_cmds.h664 struct be_queue_info *rxq, u16 cq_id,
/drivers/net/arm/
Dixp4xx_eth.c475 qmgr_disable_irq(port->plat->rxq); in eth_rx_irq()
483 unsigned int rxq = port->plat->rxq, rxfreeq = RXFREE_QUEUE(port->id); in eth_poll() local
499 if ((n = queue_get_desc(rxq, port, 0)) < 0) { in eth_poll()
505 qmgr_enable_irq(rxq); in eth_poll()
506 if (!qmgr_stat_empty(rxq) && in eth_poll()
513 qmgr_disable_irq(rxq); in eth_poll()
812 err = qmgr_request_queue(port->plat->rxq, RX_DESCS, 0, 0, in request_queues()
841 qmgr_release_queue(port->plat->rxq); in request_queues()
852 qmgr_release_queue(port->plat->rxq); in release_queues()
963 msg.byte5 = port->plat->rxq | 0x80; in eth_open()
[all …]
/drivers/net/wireless/ipw2x00/
Dipw2200.c3346 struct ipw_rx_queue *rxq) in ipw_rx_queue_reset() argument
3351 spin_lock_irqsave(&rxq->lock, flags); in ipw_rx_queue_reset()
3353 INIT_LIST_HEAD(&rxq->rx_free); in ipw_rx_queue_reset()
3354 INIT_LIST_HEAD(&rxq->rx_used); in ipw_rx_queue_reset()
3360 if (rxq->pool[i].skb != NULL) { in ipw_rx_queue_reset()
3361 pci_unmap_single(priv->pci_dev, rxq->pool[i].dma_addr, in ipw_rx_queue_reset()
3363 dev_kfree_skb(rxq->pool[i].skb); in ipw_rx_queue_reset()
3364 rxq->pool[i].skb = NULL; in ipw_rx_queue_reset()
3366 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in ipw_rx_queue_reset()
3371 rxq->read = rxq->write = 0; in ipw_rx_queue_reset()
[all …]
Dipw2100.c2603 struct ipw2100_bd_queue *rxq = &priv->rx_queue; in __ipw2100_rx_process() local
2616 if (r >= rxq->entries) { in __ipw2100_rx_process()
2621 i = (rxq->next + 1) % rxq->entries; in __ipw2100_rx_process()
2701 rxq->drv[i].status.info.field = 0; in __ipw2100_rx_process()
2703 i = (i + 1) % rxq->entries; in __ipw2100_rx_process()
2708 rxq->next = (i ? i : rxq->entries) - 1; in __ipw2100_rx_process()
2711 IPW_MEM_HOST_SHARED_RX_WRITE_INDEX, rxq->next); in __ipw2100_rx_process()
/drivers/net/wan/
Dixp4xx_hss.c632 unsigned int rxq = queue_ids[port->id].rx; in hss_hdlc_poll() local
649 if ((n = queue_get_desc(rxq, port, 0)) < 0) { in hss_hdlc_poll()
655 qmgr_enable_irq(rxq); in hss_hdlc_poll()
656 if (!qmgr_stat_empty(rxq) && in hss_hdlc_poll()
663 qmgr_disable_irq(rxq); in hss_hdlc_poll()
/drivers/infiniband/hw/cxgb3/
Diwch_cm.c108 static struct sk_buff_head rxq; variable
277 while ((skb = skb_dequeue(&rxq))) { in process_work()
2115 skb_queue_tail(&rxq, skb); in sched()
2133 skb_queue_head_init(&rxq); in iwch_cm_init()
/drivers/char/
Distallion.c2105 rp = &((cdkasy_t __iomem *) EBRDGETMEMPTR(brdp, portp->addr))->rxq; in stli_read()
2136 rp = &((cdkasy_t __iomem *) EBRDGETMEMPTR(brdp, portp->addr))->rxq; in stli_read()
3470 portp->rxsize = readw(&ap->rxq.size); in stli_startbrd()
3472 portp->rxoffset = readl(&ap->rxq.offset); in stli_startbrd()

12