/drivers/net/wireless/iwlwifi/ |
D | iwl-rx.c | 191 struct iwl_rx_queue *rxq = &priv->rxq; in iwl_rx_queue_restock() local 198 spin_lock_irqsave(&rxq->lock, flags); in iwl_rx_queue_restock() 199 write = rxq->write & ~0x7; in iwl_rx_queue_restock() 200 while ((iwl_rx_queue_space(rxq) > 0) && (rxq->free_count)) { in iwl_rx_queue_restock() 202 element = rxq->rx_free.next; in iwl_rx_queue_restock() 207 rxq->bd[rxq->write] = iwl_dma_addr2rbd_ptr(priv, rxb->aligned_dma_addr); in iwl_rx_queue_restock() 208 rxq->queue[rxq->write] = rxb; in iwl_rx_queue_restock() 209 rxq->write = (rxq->write + 1) & RX_QUEUE_MASK; in iwl_rx_queue_restock() 210 rxq->free_count--; in iwl_rx_queue_restock() 212 spin_unlock_irqrestore(&rxq->lock, flags); in iwl_rx_queue_restock() [all …]
|
D | iwl3945-base.c | 3521 struct iwl3945_rx_queue *rxq = &priv->rxq; in iwl3945_rx_queue_restock() local 3527 spin_lock_irqsave(&rxq->lock, flags); in iwl3945_rx_queue_restock() 3528 write = rxq->write & ~0x7; in iwl3945_rx_queue_restock() 3529 while ((iwl3945_rx_queue_space(rxq) > 0) && (rxq->free_count)) { in iwl3945_rx_queue_restock() 3531 element = rxq->rx_free.next; in iwl3945_rx_queue_restock() 3536 rxq->bd[rxq->write] = iwl3945_dma_addr2rbd_ptr(priv, rxb->dma_addr); in iwl3945_rx_queue_restock() 3537 rxq->queue[rxq->write] = rxb; in iwl3945_rx_queue_restock() 3538 rxq->write = (rxq->write + 1) & RX_QUEUE_MASK; in iwl3945_rx_queue_restock() 3539 rxq->free_count--; in iwl3945_rx_queue_restock() 3541 spin_unlock_irqrestore(&rxq->lock, flags); in iwl3945_rx_queue_restock() [all …]
|
D | iwl-core.h | 230 void iwl_rx_queue_free(struct iwl_priv *priv, struct iwl_rx_queue *rxq); 235 void iwl_rx_queue_reset(struct iwl_priv *priv, struct iwl_rx_queue *rxq); 237 int iwl_rx_init(struct iwl_priv *priv, struct iwl_rx_queue *rxq);
|
D | iwl-core.c | 196 struct iwl_rx_queue *rxq = &priv->rxq; in iwl_hw_nic_init() local 210 if (!rxq->bd) { in iwl_hw_nic_init() 217 iwl_rx_queue_reset(priv, rxq); in iwl_hw_nic_init() 221 iwl_rx_init(priv, rxq); in iwl_hw_nic_init() 225 rxq->need_update = 1; in iwl_hw_nic_init() 226 iwl_rx_queue_update_write_ptr(priv, rxq); in iwl_hw_nic_init()
|
D | iwl-agn.c | 1124 struct iwl_rx_queue *rxq = &priv->rxq; in iwl_rx_handle() local 1133 r = le16_to_cpu(rxq->rb_stts->closed_rb_num) & 0x0FFF; in iwl_rx_handle() 1134 i = rxq->read; in iwl_rx_handle() 1140 if (iwl_rx_queue_space(rxq) > (RX_QUEUE_SIZE / 2)) in iwl_rx_handle() 1144 rxb = rxq->queue[i]; in iwl_rx_handle() 1151 rxq->queue[i] = NULL; in iwl_rx_handle() 1211 spin_lock_irqsave(&rxq->lock, flags); in iwl_rx_handle() 1212 list_add_tail(&rxb->list, &priv->rxq.rx_used); in iwl_rx_handle() 1213 spin_unlock_irqrestore(&rxq->lock, flags); in iwl_rx_handle() 1220 priv->rxq.read = i; in iwl_rx_handle() [all …]
|
D | iwl-3945.c | 975 static int iwl3945_rx_init(struct iwl3945_priv *priv, struct iwl3945_rx_queue *rxq) in iwl3945_rx_init() argument 987 iwl3945_write_direct32(priv, FH_RCSR_RBD_BASE(0), rxq->dma_addr); in iwl3945_rx_init() 1096 struct iwl3945_rx_queue *rxq = &priv->rxq; in iwl3945_hw_nic_init() local 1188 if (!rxq->bd) { in iwl3945_hw_nic_init() 1195 iwl3945_rx_queue_reset(priv, rxq); in iwl3945_hw_nic_init() 1199 iwl3945_rx_init(priv, rxq); in iwl3945_hw_nic_init() 1213 iwl3945_write_direct32(priv, FH_RCSR_WPTR(0), rxq->write & ~7); in iwl3945_hw_nic_init()
|
D | iwl-3945.h | 575 struct iwl3945_rx_queue *rxq); 817 struct iwl3945_rx_queue rxq; member
|
D | iwl-dev.h | 868 struct iwl_rx_queue rxq; member
|
/drivers/atm/ |
D | ambassador.c | 687 amb_rxq * rxq = &dev->rxq[pool]; in rx_give() local 692 spin_lock_irqsave (&rxq->lock, flags); in rx_give() 694 if (rxq->pending < rxq->maximum) { in rx_give() 695 PRINTD (DBG_RX, "RX in slot %p", rxq->in.ptr); in rx_give() 697 *rxq->in.ptr = *rx; in rx_give() 698 rxq->pending++; in rx_give() 699 rxq->in.ptr = NEXTQ (rxq->in.ptr, rxq->in.start, rxq->in.limit); in rx_give() 701 wr_mem (dev, offsetof(amb_mem, mb.adapter.rx_address[pool]), virt_to_bus (rxq->in.ptr)); in rx_give() 703 spin_unlock_irqrestore (&rxq->lock, flags); in rx_give() 706 spin_unlock_irqrestore (&rxq->lock, flags); in rx_give() [all …]
|
D | fore200e.c | 1179 struct host_rxq* rxq = &fore200e->host_rxq; in fore200e_rx_irq() local 1186 entry = &rxq->host_entry[ rxq->head ]; in fore200e_rx_irq() 1217 FORE200E_NEXT_ENTRY(rxq->head, QUEUE_SIZE_RX); in fore200e_rx_irq() 2215 struct host_rxq* rxq = &fore200e->host_rxq; in fore200e_init_rx_queue() local 2223 &rxq->status, in fore200e_init_rx_queue() 2232 &rxq->rpd, in fore200e_init_rx_queue() 2237 fore200e->bus->dma_chunk_free(fore200e, &rxq->status); in fore200e_init_rx_queue() 2247 rxq->host_entry[ i ].status = in fore200e_init_rx_queue() 2248 FORE200E_INDEX(rxq->status.align_addr, enum status, i); in fore200e_init_rx_queue() 2249 rxq->host_entry[ i ].rpd = in fore200e_init_rx_queue() [all …]
|
D | ambassador.h | 639 amb_rxq rxq[NUM_RX_POOLS]; member
|
/drivers/net/ |
D | mv643xx_eth.c | 393 struct rx_queue rxq[8]; member 429 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) in rxq_to_mp() argument 431 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp() 439 static void rxq_enable(struct rx_queue *rxq) in rxq_enable() argument 441 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_enable() 442 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); in rxq_enable() 445 static void rxq_disable(struct rx_queue *rxq) in rxq_disable() argument 447 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_disable() 448 u8 mask = 1 << rxq->index; in rxq_disable() 496 static int rxq_process(struct rx_queue *rxq, int budget) in rxq_process() argument [all …]
|
D | ibmveth.h | 49 #define h_register_logical_lan(ua, buflst, rxq, fltlst, mac) \ argument 50 plpar_hcall_norets(H_REGISTER_LOGICAL_LAN, ua, buflst, rxq, fltlst, mac)
|
D | sky2.c | 1132 unsigned rxq = rxqaddr[sky2->port]; in sky2_rx_stop() local 1136 sky2_write8(hw, RB_ADDR(rxq, RB_CTRL), RB_DIS_OP_MD); in sky2_rx_stop() 1139 if (sky2_read8(hw, RB_ADDR(rxq, Q_RSL)) in sky2_rx_stop() 1140 == sky2_read8(hw, RB_ADDR(rxq, Q_RL))) in sky2_rx_stop() 1146 sky2_write32(hw, Q_ADDR(rxq, Q_CSR), BMU_RST_SET | BMU_FIFO_RST); in sky2_rx_stop() 1149 sky2_write32(hw, Y2_QADDR(rxq, PREF_UNIT_CTRL), PREF_UNIT_RST_SET); in sky2_rx_stop() 1289 static inline void sky2_rx_update(struct sky2_port *sky2, unsigned rxq) in sky2_rx_update() argument 1291 sky2_put_idx(sky2->hw, rxq, sky2->rx_put); in sky2_rx_update() 1307 unsigned rxq = rxqaddr[sky2->port]; in sky2_rx_start() local 1311 sky2_qset(hw, rxq); in sky2_rx_start() [all …]
|
D | xen-netfront.c | 811 struct sk_buff_head *rxq) in handle_incoming_queue() argument 816 while ((skb = __skb_dequeue(rxq)) != NULL) { in handle_incoming_queue() 859 struct sk_buff_head rxq; in xennet_poll() local 868 skb_queue_head_init(&rxq); in xennet_poll() 956 __skb_queue_tail(&rxq, skb); in xennet_poll() 964 work_done -= handle_incoming_queue(dev, &rxq); in xennet_poll()
|
/drivers/net/usb/ |
D | usbnet.c | 323 spin_lock_irqsave (&dev->rxq.lock, lockflags); in rx_submit() 346 __skb_queue_tail (&dev->rxq, skb); in rx_submit() 353 spin_unlock_irqrestore (&dev->rxq.lock, lockflags); in rx_submit() 455 defer_bh(dev, skb, &dev->rxq); in rx_complete() 541 (void) unlink_urbs (dev, &dev->rxq); in usbnet_unlink_rx_urbs() 569 temp = unlink_urbs (dev, &dev->txq) + unlink_urbs (dev, &dev->rxq); in usbnet_stop() 572 while (!skb_queue_empty(&dev->rxq) in usbnet_stop() 812 unlink_urbs (dev, &dev->rxq); in kevent() 1037 if ((dev->txq.qlen + dev->rxq.qlen + dev->done.qlen) == 0) { in usbnet_bh() 1046 int temp = dev->rxq.qlen; in usbnet_bh() [all …]
|
/drivers/net/benet/ |
D | be_main.c | 602 struct be_queue_info *rxq = &adapter->rx_obj.q; in get_rx_page_info() local 611 atomic_dec(&rxq->used); in get_rx_page_info() 619 struct be_queue_info *rxq = &adapter->rx_obj.q; in be_rx_compl_discard() local 630 index_inc(&rxq_idx, rxq->len); in be_rx_compl_discard() 641 struct be_queue_info *rxq = &adapter->rx_obj.q; in skb_fill_rx_data() local 687 index_inc(&rxq_idx, rxq->len); in skb_fill_rx_data() 762 struct be_queue_info *rxq = &adapter->rx_obj.q; in be_rx_compl_process_lro() local 782 index_inc(&rxq_idx, rxq->len); in be_rx_compl_process_lro() 838 struct be_queue_info *rxq = &adapter->rx_obj.q; in be_post_rx_frags() local 845 page_info = &page_info_tbl[rxq->head]; in be_post_rx_frags() [all …]
|
D | be_cmds.c | 455 struct be_queue_info *rxq, u16 cq_id, u16 frag_size, in be_cmd_rxq_create() argument 460 struct be_dma_mem *q_mem = &rxq->dma_mem; in be_cmd_rxq_create() 482 rxq->id = le16_to_cpu(resp->id); in be_cmd_rxq_create() 483 rxq->created = true; in be_cmd_rxq_create()
|
D | be_cmds.h | 664 struct be_queue_info *rxq, u16 cq_id,
|
/drivers/net/arm/ |
D | ixp4xx_eth.c | 475 qmgr_disable_irq(port->plat->rxq); in eth_rx_irq() 483 unsigned int rxq = port->plat->rxq, rxfreeq = RXFREE_QUEUE(port->id); in eth_poll() local 499 if ((n = queue_get_desc(rxq, port, 0)) < 0) { in eth_poll() 505 qmgr_enable_irq(rxq); in eth_poll() 506 if (!qmgr_stat_empty(rxq) && in eth_poll() 513 qmgr_disable_irq(rxq); in eth_poll() 812 err = qmgr_request_queue(port->plat->rxq, RX_DESCS, 0, 0, in request_queues() 841 qmgr_release_queue(port->plat->rxq); in request_queues() 852 qmgr_release_queue(port->plat->rxq); in release_queues() 963 msg.byte5 = port->plat->rxq | 0x80; in eth_open() [all …]
|
/drivers/net/wireless/ipw2x00/ |
D | ipw2200.c | 3346 struct ipw_rx_queue *rxq) in ipw_rx_queue_reset() argument 3351 spin_lock_irqsave(&rxq->lock, flags); in ipw_rx_queue_reset() 3353 INIT_LIST_HEAD(&rxq->rx_free); in ipw_rx_queue_reset() 3354 INIT_LIST_HEAD(&rxq->rx_used); in ipw_rx_queue_reset() 3360 if (rxq->pool[i].skb != NULL) { in ipw_rx_queue_reset() 3361 pci_unmap_single(priv->pci_dev, rxq->pool[i].dma_addr, in ipw_rx_queue_reset() 3363 dev_kfree_skb(rxq->pool[i].skb); in ipw_rx_queue_reset() 3364 rxq->pool[i].skb = NULL; in ipw_rx_queue_reset() 3366 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in ipw_rx_queue_reset() 3371 rxq->read = rxq->write = 0; in ipw_rx_queue_reset() [all …]
|
D | ipw2100.c | 2603 struct ipw2100_bd_queue *rxq = &priv->rx_queue; in __ipw2100_rx_process() local 2616 if (r >= rxq->entries) { in __ipw2100_rx_process() 2621 i = (rxq->next + 1) % rxq->entries; in __ipw2100_rx_process() 2701 rxq->drv[i].status.info.field = 0; in __ipw2100_rx_process() 2703 i = (i + 1) % rxq->entries; in __ipw2100_rx_process() 2708 rxq->next = (i ? i : rxq->entries) - 1; in __ipw2100_rx_process() 2711 IPW_MEM_HOST_SHARED_RX_WRITE_INDEX, rxq->next); in __ipw2100_rx_process()
|
/drivers/net/wan/ |
D | ixp4xx_hss.c | 632 unsigned int rxq = queue_ids[port->id].rx; in hss_hdlc_poll() local 649 if ((n = queue_get_desc(rxq, port, 0)) < 0) { in hss_hdlc_poll() 655 qmgr_enable_irq(rxq); in hss_hdlc_poll() 656 if (!qmgr_stat_empty(rxq) && in hss_hdlc_poll() 663 qmgr_disable_irq(rxq); in hss_hdlc_poll()
|
/drivers/infiniband/hw/cxgb3/ |
D | iwch_cm.c | 108 static struct sk_buff_head rxq; variable 277 while ((skb = skb_dequeue(&rxq))) { in process_work() 2115 skb_queue_tail(&rxq, skb); in sched() 2133 skb_queue_head_init(&rxq); in iwch_cm_init()
|
/drivers/char/ |
D | istallion.c | 2105 rp = &((cdkasy_t __iomem *) EBRDGETMEMPTR(brdp, portp->addr))->rxq; in stli_read() 2136 rp = &((cdkasy_t __iomem *) EBRDGETMEMPTR(brdp, portp->addr))->rxq; in stli_read() 3470 portp->rxsize = readw(&ap->rxq.size); in stli_startbrd() 3472 portp->rxoffset = readl(&ap->rxq.offset); in stli_startbrd()
|