Home
last modified time | relevance | path

Searched refs:iov_offset (Results 1 – 13 of 13) sorted by relevance

/kernel/linux/linux-5.10/lib/
Diov_iter.c83 size_t skip = i->iov_offset; \
105 size_t skip = i->iov_offset; \
138 i->iov_offset = skip; \
181 skip = i->iov_offset; in copy_page_to_iter_iovec()
244 i->iov_offset = skip; in copy_page_to_iter_iovec()
265 skip = i->iov_offset; in copy_page_from_iter_iovec()
328 i->iov_offset = skip; in copy_page_from_iter_iovec()
343 if (i->iov_offset) { in sanity()
351 if (unlikely(p->offset + p->len != i->iov_offset)) in sanity()
359 printk(KERN_ERR "idx = %d, offset = %zd\n", i_head, i->iov_offset); in sanity()
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/google/gve/
Dgve_tx.c95 iov[0].iov_offset = fifo->head; in gve_tx_alloc_fifo()
106 iov[1].iov_offset = 0; /* Start of fifo*/ in gve_tx_alloc_fifo()
398 u64 iov_offset, u64 iov_len) in gve_dma_sync_for_device() argument
400 u64 last_page = (iov_offset + iov_len - 1) / PAGE_SIZE; in gve_dma_sync_for_device()
401 u64 first_page = iov_offset / PAGE_SIZE; in gve_dma_sync_for_device()
449 info->iov[hdr_nfrags - 1].iov_offset); in gve_tx_add_skb()
452 tx->tx_fifo.base + info->iov[hdr_nfrags - 1].iov_offset, in gve_tx_add_skb()
455 info->iov[hdr_nfrags - 1].iov_offset, in gve_tx_add_skb()
465 info->iov[i].iov_offset); in gve_tx_add_skb()
468 tx->tx_fifo.base + info->iov[i].iov_offset, in gve_tx_add_skb()
[all …]
Dgve.h105 u32 iov_offset; /* offset into this segment */ member
/kernel/linux/linux-5.10/include/linux/
Duio.h30 size_t iov_offset; member
42 size_t iov_offset; member
67 state->iov_offset = iter->iov_offset; in iov_iter_save_state()
122 .iov_base = iter->iov->iov_base + iter->iov_offset, in iov_iter_iovec()
124 iter->iov->iov_len - iter->iov_offset), in iov_iter_iovec()
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/
Duser_sdma.c745 u64 iov_offset = *iov_offset_ptr; in user_sdma_txadd() local
750 offset = offset_in_page(base + iovec->offset + iov_offset); in user_sdma_txadd()
751 pageidx = (((iovec->offset + iov_offset + base) - (base & PAGE_MASK)) >> in user_sdma_txadd()
762 iov_offset += len; in user_sdma_txadd()
767 iovec->offset += iov_offset; in user_sdma_txadd()
769 iov_offset = 0; in user_sdma_txadd()
774 *iov_offset_ptr = iov_offset; in user_sdma_txadd()
810 u64 iov_offset = 0; in user_sdma_send_pkts() local
910 &queued, &data_sent, &iov_offset); in user_sdma_send_pkts()
923 iovec->offset += iov_offset; in user_sdma_send_pkts()
/kernel/linux/linux-5.10/block/
Dbio.c970 if (WARN_ON_ONCE(iter->iov_offset > bv->bv_len)) in __bio_iov_bvec_add_pages()
973 len = min_t(size_t, bv->bv_len - iter->iov_offset, iter->count); in __bio_iov_bvec_add_pages()
975 bv->bv_offset + iter->iov_offset); in __bio_iov_bvec_add_pages()
/kernel/linux/linux-5.10/net/9p/
Dtrans_virtio.c346 p = data->kvec->iov_base + data->iov_offset; in p9_get_mapped_pages()
/kernel/linux/linux-5.10/drivers/nvme/host/
Dtcp.c237 return req->iter.bvec->bv_offset + req->iter.iov_offset; in nvme_tcp_req_cur_offset()
248 return req->iter.iov_offset; in nvme_tcp_req_offset()
287 req->iter.iov_offset = offset; in nvme_tcp_init_iter()
/kernel/linux/linux-5.10/fs/
Dsplice.c317 to.iov_offset = 0; in generic_file_splice_read()
/kernel/linux/linux-5.10/drivers/block/
Dloop.c563 iter.iov_offset = offset; in lo_rw_aio()
/kernel/linux/linux-5.10/fs/cifs/
Dfile.c2951 from->iov_offset, from->count); in cifs_write_from_iter()
3686 direct_iov.iov_offset, in cifs_send_async_read()
/kernel/linux/linux-5.10/fs/fuse/
Dfile.c1366 return (unsigned long)ii->iov->iov_base + ii->iov_offset; in fuse_get_user_addr()
/kernel/linux/linux-5.10/io_uring/
Dio_uring.c3128 iter->iov_offset = offset & ~PAGE_MASK; in __io_import_fixed()