Lines Matching refs:off
538 loff_t off; member
550 u32 off; member
1290 return !req->timeout.off; in io_is_timeout_noseq()
2981 kiocb->ki_pos = READ_ONCE(sqe->off); in io_prep_rw()
3910 if (sqe->ioprio || sqe->off || sqe->len || sqe->buf_index || in io_unlinkat_prep()
3957 if (unlikely(sqe->ioprio || sqe->off || sqe->addr || sqe->rw_flags || in io_shutdown_prep()
4011 if (READ_ONCE(sqe->splice_off_in) || READ_ONCE(sqe->off)) in io_tee_prep()
4051 sp->off_out = READ_ONCE(sqe->off); in io_splice_prep()
4117 req->sync.off = READ_ONCE(sqe->off); in io_fsync_prep()
4124 loff_t end = req->sync.off + req->sync.len; in io_fsync()
4131 ret = vfs_fsync_range(req->file, req->sync.off, in io_fsync()
4149 req->sync.off = READ_ONCE(sqe->off); in io_fallocate_prep()
4162 ret = vfs_fallocate(req->file, req->sync.mode, req->sync.off, in io_fallocate()
4313 if (sqe->ioprio || sqe->rw_flags || sqe->addr || sqe->len || sqe->off || in io_remove_buffers_prep()
4407 tmp = READ_ONCE(sqe->off); in io_provide_buffers_prep()
4482 req->epoll.fd = READ_ONCE(sqe->off); in io_epoll_ctl_prep()
4521 if (sqe->ioprio || sqe->buf_index || sqe->off || sqe->splice_fd_in) in io_madvise_prep()
4561 req->fadvise.offset = READ_ONCE(sqe->off); in io_fadvise_prep()
4629 if (sqe->ioprio || sqe->off || sqe->addr || sqe->len || in io_close_prep()
4704 req->sync.off = READ_ONCE(sqe->off); in io_sfr_prep()
4718 ret = sync_file_range(req->file, req->sync.off, req->sync.len, in io_sync_file_range()
5969 upd->new_user_data = READ_ONCE(sqe->off); in io_poll_update_prep()
5987 if (sqe->ioprio || sqe->buf_index || sqe->off || sqe->addr) in io_poll_add_prep()
6172 req->timeout.off = 0; /* noseq */ in io_timeout_update()
6256 u32 off = READ_ONCE(sqe->off); in io_timeout_prep() local
6263 if (off && is_timeout_link) in io_timeout_prep()
6273 req->timeout.off = off; in io_timeout_prep()
6274 if (unlikely(off && !req->ctx->off_timeout_used)) in io_timeout_prep()
6309 u32 tail, off = req->timeout.off; in io_timeout() local
6324 req->timeout.target_seq = tail + off; in io_timeout()
6343 if (off >= nxt->timeout.target_seq - tail) in io_timeout()
6423 if (sqe->ioprio || sqe->off || sqe->len || sqe->cancel_flags || in io_async_cancel_prep()
6468 req->rsrc_update.offset = READ_ONCE(sqe->off); in io_rsrc_update_prep()
7932 unsigned int off = idx & IO_RSRC_TAG_TABLE_MASK; in io_get_tag_slot() local
7935 return &data->tags[table_idx][off]; in io_get_tag_slot()
8885 size_t off, sq_array_size; in rings_size() local
8887 off = struct_size(rings, cqes, cq_entries); in rings_size()
8888 if (off == SIZE_MAX) in rings_size()
8892 off = ALIGN(off, SMP_CACHE_BYTES); in rings_size()
8893 if (off == 0) in rings_size()
8898 *sq_offset = off; in rings_size()
8904 if (check_add_overflow(off, sq_array_size, &off)) in rings_size()
8907 return off; in rings_size()
9065 unsigned long off, start, end, ubuf; in io_sqe_buffer_register() local
9135 off = ubuf & ~PAGE_MASK; in io_sqe_buffer_register()
9140 vec_len = min_t(size_t, size, PAGE_SIZE - off); in io_sqe_buffer_register()
9143 imu->bvec[i].bv_offset = off; in io_sqe_buffer_register()
9144 off = 0; in io_sqe_buffer_register()
11124 BUILD_BUG_SQE_ELEM(8, __u64, off); in io_uring_init()