Searched refs:iov_offset (Results 1 – 13 of 13) sorted by relevance
/kernel/linux/linux-5.10/lib/ |
D | iov_iter.c | 83 size_t skip = i->iov_offset; \ 105 size_t skip = i->iov_offset; \ 138 i->iov_offset = skip; \ 181 skip = i->iov_offset; in copy_page_to_iter_iovec() 244 i->iov_offset = skip; in copy_page_to_iter_iovec() 265 skip = i->iov_offset; in copy_page_from_iter_iovec() 328 i->iov_offset = skip; in copy_page_from_iter_iovec() 343 if (i->iov_offset) { in sanity() 351 if (unlikely(p->offset + p->len != i->iov_offset)) in sanity() 359 printk(KERN_ERR "idx = %d, offset = %zd\n", i_head, i->iov_offset); in sanity() [all …]
|
/kernel/linux/linux-5.10/drivers/net/ethernet/google/gve/ |
D | gve_tx.c | 95 iov[0].iov_offset = fifo->head; in gve_tx_alloc_fifo() 106 iov[1].iov_offset = 0; /* Start of fifo*/ in gve_tx_alloc_fifo() 398 u64 iov_offset, u64 iov_len) in gve_dma_sync_for_device() argument 400 u64 last_page = (iov_offset + iov_len - 1) / PAGE_SIZE; in gve_dma_sync_for_device() 401 u64 first_page = iov_offset / PAGE_SIZE; in gve_dma_sync_for_device() 449 info->iov[hdr_nfrags - 1].iov_offset); in gve_tx_add_skb() 452 tx->tx_fifo.base + info->iov[hdr_nfrags - 1].iov_offset, in gve_tx_add_skb() 455 info->iov[hdr_nfrags - 1].iov_offset, in gve_tx_add_skb() 465 info->iov[i].iov_offset); in gve_tx_add_skb() 468 tx->tx_fifo.base + info->iov[i].iov_offset, in gve_tx_add_skb() [all …]
|
D | gve.h | 105 u32 iov_offset; /* offset into this segment */ member
|
/kernel/linux/linux-5.10/include/linux/ |
D | uio.h | 30 size_t iov_offset; member 42 size_t iov_offset; member 67 state->iov_offset = iter->iov_offset; in iov_iter_save_state() 122 .iov_base = iter->iov->iov_base + iter->iov_offset, in iov_iter_iovec() 124 iter->iov->iov_len - iter->iov_offset), in iov_iter_iovec()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/ |
D | user_sdma.c | 745 u64 iov_offset = *iov_offset_ptr; in user_sdma_txadd() local 750 offset = offset_in_page(base + iovec->offset + iov_offset); in user_sdma_txadd() 751 pageidx = (((iovec->offset + iov_offset + base) - (base & PAGE_MASK)) >> in user_sdma_txadd() 762 iov_offset += len; in user_sdma_txadd() 767 iovec->offset += iov_offset; in user_sdma_txadd() 769 iov_offset = 0; in user_sdma_txadd() 774 *iov_offset_ptr = iov_offset; in user_sdma_txadd() 810 u64 iov_offset = 0; in user_sdma_send_pkts() local 910 &queued, &data_sent, &iov_offset); in user_sdma_send_pkts() 923 iovec->offset += iov_offset; in user_sdma_send_pkts()
|
/kernel/linux/linux-5.10/block/ |
D | bio.c | 970 if (WARN_ON_ONCE(iter->iov_offset > bv->bv_len)) in __bio_iov_bvec_add_pages() 973 len = min_t(size_t, bv->bv_len - iter->iov_offset, iter->count); in __bio_iov_bvec_add_pages() 975 bv->bv_offset + iter->iov_offset); in __bio_iov_bvec_add_pages()
|
/kernel/linux/linux-5.10/net/9p/ |
D | trans_virtio.c | 346 p = data->kvec->iov_base + data->iov_offset; in p9_get_mapped_pages()
|
/kernel/linux/linux-5.10/drivers/nvme/host/ |
D | tcp.c | 237 return req->iter.bvec->bv_offset + req->iter.iov_offset; in nvme_tcp_req_cur_offset() 248 return req->iter.iov_offset; in nvme_tcp_req_offset() 287 req->iter.iov_offset = offset; in nvme_tcp_init_iter()
|
/kernel/linux/linux-5.10/fs/ |
D | splice.c | 317 to.iov_offset = 0; in generic_file_splice_read()
|
/kernel/linux/linux-5.10/drivers/block/ |
D | loop.c | 563 iter.iov_offset = offset; in lo_rw_aio()
|
/kernel/linux/linux-5.10/fs/cifs/ |
D | file.c | 2951 from->iov_offset, from->count); in cifs_write_from_iter() 3686 direct_iov.iov_offset, in cifs_send_async_read()
|
/kernel/linux/linux-5.10/fs/fuse/ |
D | file.c | 1366 return (unsigned long)ii->iov->iov_base + ii->iov_offset; in fuse_get_user_addr()
|
/kernel/linux/linux-5.10/io_uring/ |
D | io_uring.c | 3128 iter->iov_offset = offset & ~PAGE_MASK; in __io_import_fixed()
|