Searched refs:TX_RING_SIZE (Results 1 – 15 of 15) sorted by relevance
5 #define TX_RING_SIZE 4 macro29 struct natsemi_tx tx[TX_RING_SIZE];41 struct io_buffer *tx_iobuf[TX_RING_SIZE];
257 malloc_dma(TX_RING_SIZE * sizeof(struct TXD), TX_RING_ALIGN); in a3c90x_setup_tx_ring()264 memset(p->tx_ring, 0, TX_RING_SIZE * sizeof(struct TXD)); in a3c90x_setup_tx_ring()303 p->tx_tail = (p->tx_tail + 1) % TX_RING_SIZE; in a3c90x_process_tx_packets()312 free_dma(p->tx_ring, TX_RING_SIZE * sizeof(struct TXD)); in a3c90x_free_tx_ring()337 if (inf_3c90x->tx_cnt == TX_RING_SIZE) { in a3c90x_transmit()346 (((inf_3c90x->tx_cur + TX_RING_SIZE) - 1) % TX_RING_SIZE); in a3c90x_transmit()384 inf_3c90x->tx_cur = (inf_3c90x->tx_cur + 1) % TX_RING_SIZE; in a3c90x_transmit()
20 #define TX_RING_SIZE 2 /* use at least 2 buffers for TX */ macro93 struct epic_tx_desc tx_ring[TX_RING_SIZE]96 unsigned char tx_packet[PKT_BUF_SZ * TX_RING_SIZE];289 for (i = 0; i < TX_RING_SIZE; i++) { in epic100_init_ring()317 entry = cur_tx % TX_RING_SIZE; in epic100_transmit()
105 #define TX_RING_SIZE 2 macro227 struct w840_tx_desc tx_ring[TX_RING_SIZE];258 char tx_packet[PKT_BUF_SZ * TX_RING_SIZE];497 entry = w840private.cur_tx % TX_RING_SIZE; in w89c840_transmit()515 if (entry >= TX_RING_SIZE-1) /* Wrap ring */ in w89c840_transmit()948 for (i = 0; i < TX_RING_SIZE; i++) { in init_ring()
356 for (i = 0 ; i < TX_RING_SIZE ; i++) { in natsemi_open()357 np->tx[i].link = virt_to_bus ((i + 1 < TX_RING_SIZE) ? &np->tx[i + 1] : &np->tx[0]); in natsemi_open()485 np->tx_cur = (np->tx_cur + 1) % TX_RING_SIZE; in natsemi_transmit()540 np->tx_dirty = (np->tx_dirty + 1) % TX_RING_SIZE; in natsemi_poll()541 i = (i + 1) % TX_RING_SIZE; in natsemi_poll()
258 #define TX_RING_SIZE 8 macro297 struct io_buffer *tx_iobuf[TX_RING_SIZE];
421 #define TX_RING_SIZE 2 macro550 char txbuf[TX_RING_SIZE * PKT_BUF_SZ + 32];552 char txdesc[TX_RING_SIZE * sizeof (struct rhine_tx_desc) + 32];665 char *tx_buffs[TX_RING_SIZE];675 struct sk_buff *tx_skbuff[TX_RING_SIZE];740 for (i = 0; i < TX_RING_SIZE; i++) in rhine_init_ring()1244 for (i = 0; i < TX_RING_SIZE; i++) in rhine_reset()1361 entry = tp->cur_tx % TX_RING_SIZE; in rhine_transmit()
152 #define TX_RING_SIZE (1 << (PCNET32_LOG_TX_BUFFERS)) macro153 #define TX_RING_MOD_MASK (TX_RING_SIZE - 1)217 struct pcnet32_tx_head tx_ring[TX_RING_SIZE]221 unsigned char txb[TX_RING_SIZE][PKT_BUF_SZ];236 struct sk_buff *tx_skbuff[TX_RING_SIZE];394 for (i = 0; i < TX_RING_SIZE; i++) { in pcnet32_init_ring()
88 #define TX_RING_SIZE 4 macro92 struct io_buffer *iobuf[TX_RING_SIZE];397 rtl->tx.next = ( rtl->tx.next + 1 ) % TX_RING_SIZE; in rtl_transmit()425 for ( i = 0 ; i < TX_RING_SIZE ; i++ ) { in rtl_poll()
65 #define TX_RING_SIZE 128 macro66 #define TX_PENDING (TX_RING_SIZE - 1)871 sky2->tx_prod = RING_NEXT(sky2->tx_prod, TX_RING_SIZE); in get_tx_le()1117 free_dma(sky2->tx_le, TX_RING_SIZE * sizeof(struct sky2_tx_le)); in sky2_free_rings()1139 sky2->tx_le = malloc_dma(TX_RING_SIZE * sizeof(struct sky2_tx_le), TX_RING_ALIGN); in sky2_up()1143 memset(sky2->tx_le, 0, TX_RING_SIZE * sizeof(struct sky2_tx_le)); in sky2_up()1145 sky2->tx_ring = zalloc(TX_RING_SIZE * sizeof(struct tx_ring_info)); in sky2_up()1195 TX_RING_SIZE - 1); in sky2_up()1221 return (head - tail) & (TX_RING_SIZE - 1); in tx_dist()1285 assert(done < TX_RING_SIZE); in sky2_tx_complete()[all …]
85 #define TX_RING_SIZE 16 macro222 struct boom_tx_desc tx_ring[TX_RING_SIZE];225 struct sk_buff *tx_skbuff[TX_RING_SIZE];390 for (i = 0; i < TX_RING_SIZE; i++) in t515_reset()
50 #define TX_RING_SIZE 2 macro306 u8 txb[PKT_BUF_SZ * TX_RING_SIZE] __attribute__ ((aligned(8)));328 struct mtd_desc tx_ring[TX_RING_SIZE];
91 #define TX_RING_SIZE 2 macro228 static struct netdev_desc tx_ring[TX_RING_SIZE];236 unsigned char txb[PKT_BUF_SZ * TX_RING_SIZE];
30 #define TX_RING_SIZE ( NUM_TX_DESC * sizeof ( struct skge_rx_desc ) ) macro32 #define RING_SIZE ( TX_RING_SIZE + RX_RING_SIZE )
433 #define TX_RING_SIZE 2 macro436 struct tulip_tx_desc tx_ring[TX_RING_SIZE];