Lines Matching refs:rx_buf
79 efx_rx_buf_next(struct efx_rx_queue *rx_queue, struct efx_rx_buffer *rx_buf) in efx_rx_buf_next() argument
81 if (unlikely(rx_buf == efx_rx_buffer(rx_queue, rx_queue->ptr_mask))) in efx_rx_buf_next()
84 return rx_buf + 1; in efx_rx_buf_next()
88 struct efx_rx_buffer *rx_buf, in efx_sync_rx_buffer() argument
91 dma_sync_single_for_cpu(&efx->pci_dev->dev, rx_buf->dma_addr, len, in efx_sync_rx_buffer()
155 struct efx_rx_buffer *rx_buf; in efx_init_rx_buffers() local
192 rx_buf = efx_rx_buffer(rx_queue, index); in efx_init_rx_buffers()
193 rx_buf->dma_addr = dma_addr + efx->rx_ip_align; in efx_init_rx_buffers()
194 rx_buf->page = page; in efx_init_rx_buffers()
195 rx_buf->page_offset = page_offset + efx->rx_ip_align; in efx_init_rx_buffers()
196 rx_buf->len = efx->rx_dma_len; in efx_init_rx_buffers()
197 rx_buf->flags = 0; in efx_init_rx_buffers()
204 rx_buf->flags = EFX_RX_BUF_LAST_IN_PAGE; in efx_init_rx_buffers()
214 struct efx_rx_buffer *rx_buf) in efx_unmap_rx_buffer() argument
216 struct page *page = rx_buf->page; in efx_unmap_rx_buffer()
228 struct efx_rx_buffer *rx_buf, in efx_free_rx_buffers() argument
232 if (rx_buf->page) { in efx_free_rx_buffers()
233 put_page(rx_buf->page); in efx_free_rx_buffers()
234 rx_buf->page = NULL; in efx_free_rx_buffers()
236 rx_buf = efx_rx_buf_next(rx_queue, rx_buf); in efx_free_rx_buffers()
245 struct efx_rx_buffer *rx_buf) in efx_recycle_rx_page() argument
247 struct page *page = rx_buf->page; in efx_recycle_rx_page()
253 if (!(rx_buf->flags & EFX_RX_BUF_LAST_IN_PAGE)) in efx_recycle_rx_page()
272 efx_unmap_rx_buffer(efx, rx_buf); in efx_recycle_rx_page()
273 put_page(rx_buf->page); in efx_recycle_rx_page()
277 struct efx_rx_buffer *rx_buf) in efx_fini_rx_buffer() argument
280 if (rx_buf->page) in efx_fini_rx_buffer()
281 put_page(rx_buf->page); in efx_fini_rx_buffer()
284 if (rx_buf->flags & EFX_RX_BUF_LAST_IN_PAGE) { in efx_fini_rx_buffer()
285 efx_unmap_rx_buffer(rx_queue->efx, rx_buf); in efx_fini_rx_buffer()
286 efx_free_rx_buffers(rx_queue, rx_buf, 1); in efx_fini_rx_buffer()
288 rx_buf->page = NULL; in efx_fini_rx_buffer()
293 struct efx_rx_buffer *rx_buf, in efx_recycle_rx_pages() argument
299 efx_recycle_rx_page(channel, rx_buf); in efx_recycle_rx_pages()
300 rx_buf = efx_rx_buf_next(rx_queue, rx_buf); in efx_recycle_rx_pages()
305 struct efx_rx_buffer *rx_buf, in efx_discard_rx_packet() argument
310 efx_recycle_rx_pages(channel, rx_buf, n_frags); in efx_discard_rx_packet()
312 efx_free_rx_buffers(rx_queue, rx_buf, n_frags); in efx_discard_rx_packet()
389 struct efx_rx_buffer *rx_buf, in efx_rx_packet__check_len() argument
393 unsigned max_len = rx_buf->len - efx->type->rx_buffer_padding; in efx_rx_packet__check_len()
401 rx_buf->flags |= EFX_RX_PKT_DISCARD; in efx_rx_packet__check_len()
403 if ((len > rx_buf->len) && EFX_WORKAROUND_8071(efx)) { in efx_rx_packet__check_len()
426 efx_rx_packet_gro(struct efx_channel *channel, struct efx_rx_buffer *rx_buf, in efx_rx_packet_gro() argument
439 efx_free_rx_buffers(rx_queue, rx_buf, n_frags); in efx_rx_packet_gro()
446 skb->ip_summed = ((rx_buf->flags & EFX_RX_PKT_CSUMMED) ? in efx_rx_packet_gro()
451 rx_buf->page, rx_buf->page_offset, in efx_rx_packet_gro()
452 rx_buf->len); in efx_rx_packet_gro()
453 rx_buf->page = NULL; in efx_rx_packet_gro()
454 skb->len += rx_buf->len; in efx_rx_packet_gro()
458 rx_buf = efx_rx_buf_next(&channel->rx_queue, rx_buf); in efx_rx_packet_gro()
474 struct efx_rx_buffer *rx_buf, in efx_rx_mk_skb() argument
490 EFX_BUG_ON_PARANOID(rx_buf->len < hdr_len); in efx_rx_mk_skb()
498 if (rx_buf->len > hdr_len) { in efx_rx_mk_skb()
499 rx_buf->page_offset += hdr_len; in efx_rx_mk_skb()
500 rx_buf->len -= hdr_len; in efx_rx_mk_skb()
504 rx_buf->page, rx_buf->page_offset, in efx_rx_mk_skb()
505 rx_buf->len); in efx_rx_mk_skb()
506 rx_buf->page = NULL; in efx_rx_mk_skb()
507 skb->len += rx_buf->len; in efx_rx_mk_skb()
508 skb->data_len += rx_buf->len; in efx_rx_mk_skb()
512 rx_buf = efx_rx_buf_next(&channel->rx_queue, rx_buf); in efx_rx_mk_skb()
515 __free_pages(rx_buf->page, efx->rx_buffer_order); in efx_rx_mk_skb()
516 rx_buf->page = NULL; in efx_rx_mk_skb()
535 struct efx_rx_buffer *rx_buf; in efx_rx_packet() local
539 rx_buf = efx_rx_buffer(rx_queue, index); in efx_rx_packet()
540 rx_buf->flags |= flags; in efx_rx_packet()
545 efx_rx_packet__check_len(rx_queue, rx_buf, len); in efx_rx_packet()
553 WARN_ON(!(len == 0 && rx_buf->flags & EFX_RX_PKT_DISCARD)); in efx_rx_packet()
554 rx_buf->flags |= EFX_RX_PKT_DISCARD; in efx_rx_packet()
561 (rx_buf->flags & EFX_RX_PKT_CSUMMED) ? " [SUMMED]" : "", in efx_rx_packet()
562 (rx_buf->flags & EFX_RX_PKT_DISCARD) ? " [DISCARD]" : ""); in efx_rx_packet()
567 if (unlikely(rx_buf->flags & EFX_RX_PKT_DISCARD)) { in efx_rx_packet()
569 efx_discard_rx_packet(channel, rx_buf, n_frags); in efx_rx_packet()
574 rx_buf->len = len; in efx_rx_packet()
579 efx_sync_rx_buffer(efx, rx_buf, rx_buf->len); in efx_rx_packet()
584 prefetch(efx_rx_buf_va(rx_buf)); in efx_rx_packet()
586 rx_buf->page_offset += efx->rx_prefix_size; in efx_rx_packet()
587 rx_buf->len -= efx->rx_prefix_size; in efx_rx_packet()
596 rx_buf = efx_rx_buf_next(rx_queue, rx_buf); in efx_rx_packet()
599 efx_sync_rx_buffer(efx, rx_buf, efx->rx_dma_len); in efx_rx_packet()
601 rx_buf->len = len - (n_frags - 1) * efx->rx_dma_len; in efx_rx_packet()
602 efx_sync_rx_buffer(efx, rx_buf, rx_buf->len); in efx_rx_packet()
606 rx_buf = efx_rx_buffer(rx_queue, index); in efx_rx_packet()
607 efx_recycle_rx_pages(channel, rx_buf, n_frags); in efx_rx_packet()
618 struct efx_rx_buffer *rx_buf, in efx_rx_deliver() argument
622 u16 hdr_len = min_t(u16, rx_buf->len, EFX_SKB_HEADERS); in efx_rx_deliver()
624 skb = efx_rx_mk_skb(channel, rx_buf, n_frags, eh, hdr_len); in efx_rx_deliver()
629 efx_free_rx_buffers(rx_queue, rx_buf, n_frags); in efx_rx_deliver()
636 if (likely(rx_buf->flags & EFX_RX_PKT_CSUMMED)) in efx_rx_deliver()
653 struct efx_rx_buffer *rx_buf = in __efx_rx_packet() local
655 u8 *eh = efx_rx_buf_va(rx_buf); in __efx_rx_packet()
660 if (rx_buf->flags & EFX_RX_PKT_PREFIX_LEN) in __efx_rx_packet()
661 rx_buf->len = le16_to_cpup((__le16 *) in __efx_rx_packet()
670 efx_loopback_rx_packet(efx, eh, rx_buf->len); in __efx_rx_packet()
672 efx_free_rx_buffers(rx_queue, rx_buf, in __efx_rx_packet()
678 rx_buf->flags &= ~EFX_RX_PKT_CSUMMED; in __efx_rx_packet()
680 if ((rx_buf->flags & EFX_RX_PKT_TCP) && !channel->type->receive_skb && in __efx_rx_packet()
682 efx_rx_packet_gro(channel, rx_buf, channel->rx_pkt_n_frags, eh); in __efx_rx_packet()
684 efx_rx_deliver(channel, eh, rx_buf, channel->rx_pkt_n_frags); in __efx_rx_packet()
787 struct efx_rx_buffer *rx_buf; in efx_fini_rx_queue() local
799 rx_buf = efx_rx_buffer(rx_queue, index); in efx_fini_rx_queue()
800 efx_fini_rx_buffer(rx_queue, rx_buf); in efx_fini_rx_queue()