• Home
  • Raw
  • Download

Lines Matching refs:tx_cb

1919 	struct ql_tx_buf_cb *tx_cb;  in ql_process_mac_tx_intr()  local
1927 tx_cb = &qdev->tx_buf[mac_rsp->transaction_id]; in ql_process_mac_tx_intr()
1938 if (tx_cb->seg_count == 0) { in ql_process_mac_tx_intr()
1947 dma_unmap_addr(&tx_cb->map[0], mapaddr), in ql_process_mac_tx_intr()
1948 dma_unmap_len(&tx_cb->map[0], maplen), in ql_process_mac_tx_intr()
1950 tx_cb->seg_count--; in ql_process_mac_tx_intr()
1951 if (tx_cb->seg_count) { in ql_process_mac_tx_intr()
1952 for (i = 1; i < tx_cb->seg_count; i++) { in ql_process_mac_tx_intr()
1954 dma_unmap_addr(&tx_cb->map[i], in ql_process_mac_tx_intr()
1956 dma_unmap_len(&tx_cb->map[i], maplen), in ql_process_mac_tx_intr()
1961 qdev->ndev->stats.tx_bytes += tx_cb->skb->len; in ql_process_mac_tx_intr()
1964 dev_kfree_skb_irq(tx_cb->skb); in ql_process_mac_tx_intr()
1965 tx_cb->skb = NULL; in ql_process_mac_tx_intr()
2306 struct ql_tx_buf_cb *tx_cb, in ql_send_map() argument
2318 seg_cnt = tx_cb->seg_count; in ql_send_map()
2336 dma_unmap_addr_set(&tx_cb->map[seg], mapaddr, map); in ql_send_map()
2337 dma_unmap_len_set(&tx_cb->map[seg], maplen, len); in ql_send_map()
2345 oal = tx_cb->oal; in ql_send_map()
2376 dma_unmap_addr_set(&tx_cb->map[seg], mapaddr, map); in ql_send_map()
2377 dma_unmap_len_set(&tx_cb->map[seg], maplen, in ql_send_map()
2398 dma_unmap_addr_set(&tx_cb->map[seg], mapaddr, map); in ql_send_map()
2399 dma_unmap_len_set(&tx_cb->map[seg], maplen, skb_frag_size(frag)); in ql_send_map()
2413 oal = tx_cb->oal; in ql_send_map()
2427 dma_unmap_addr(&tx_cb->map[seg], mapaddr), in ql_send_map()
2428 dma_unmap_len(&tx_cb->map[seg], maplen), in ql_send_map()
2435 dma_unmap_addr(&tx_cb->map[seg], mapaddr), in ql_send_map()
2436 dma_unmap_len(&tx_cb->map[seg], maplen), in ql_send_map()
2441 dma_unmap_addr(&tx_cb->map[0], mapaddr), in ql_send_map()
2442 dma_unmap_addr(&tx_cb->map[0], maplen), in ql_send_map()
2466 struct ql_tx_buf_cb *tx_cb; in ql3xxx_send() local
2473 tx_cb = &qdev->tx_buf[qdev->req_producer_index]; in ql3xxx_send()
2474 tx_cb->seg_count = ql_get_seg_count(qdev, in ql3xxx_send()
2476 if (tx_cb->seg_count == -1) { in ql3xxx_send()
2481 mac_iocb_ptr = tx_cb->queue_entry; in ql3xxx_send()
2488 tx_cb->skb = skb; in ql3xxx_send()
2493 if (ql_send_map(qdev, mac_iocb_ptr, tx_cb, skb) != NETDEV_TX_OK) { in ql3xxx_send()
2809 struct ql_tx_buf_cb *tx_cb; in ql_free_send_free_list() local
2812 tx_cb = &qdev->tx_buf[0]; in ql_free_send_free_list()
2814 kfree(tx_cb->oal); in ql_free_send_free_list()
2815 tx_cb->oal = NULL; in ql_free_send_free_list()
2816 tx_cb++; in ql_free_send_free_list()
2822 struct ql_tx_buf_cb *tx_cb; in ql_create_send_free_list() local
2829 tx_cb = &qdev->tx_buf[i]; in ql_create_send_free_list()
2830 tx_cb->skb = NULL; in ql_create_send_free_list()
2831 tx_cb->queue_entry = req_q_curr; in ql_create_send_free_list()
2833 tx_cb->oal = kmalloc(512, GFP_KERNEL); in ql_create_send_free_list()
2834 if (tx_cb->oal == NULL) in ql_create_send_free_list()
3625 struct ql_tx_buf_cb *tx_cb; in ql_reset_work() local
3639 tx_cb = &qdev->tx_buf[i]; in ql_reset_work()
3640 if (tx_cb->skb) { in ql_reset_work()
3644 dma_unmap_addr(&tx_cb->map[0], in ql_reset_work()
3646 dma_unmap_len(&tx_cb->map[0], maplen), in ql_reset_work()
3648 for (j = 1; j < tx_cb->seg_count; j++) { in ql_reset_work()
3650 dma_unmap_addr(&tx_cb->map[j], in ql_reset_work()
3652 dma_unmap_len(&tx_cb->map[j], in ql_reset_work()
3656 dev_kfree_skb(tx_cb->skb); in ql_reset_work()
3657 tx_cb->skb = NULL; in ql_reset_work()