Lines Matching refs:off
528 loff_t off; member
540 u32 off; member
1283 return !req->timeout.off; in io_is_timeout_noseq()
2976 kiocb->ki_pos = READ_ONCE(sqe->off); in io_prep_rw()
3905 if (sqe->ioprio || sqe->off || sqe->len || sqe->buf_index || in io_unlinkat_prep()
3954 if (sqe->ioprio || sqe->off || sqe->rw_flags || sqe->buf_index || in io_mkdirat_prep()
4095 if (unlikely(sqe->ioprio || sqe->off || sqe->addr || sqe->rw_flags || in io_shutdown_prep()
4149 if (READ_ONCE(sqe->splice_off_in) || READ_ONCE(sqe->off)) in io_tee_prep()
4189 sp->off_out = READ_ONCE(sqe->off); in io_splice_prep()
4255 req->sync.off = READ_ONCE(sqe->off); in io_fsync_prep()
4262 loff_t end = req->sync.off + req->sync.len; in io_fsync()
4269 ret = vfs_fsync_range(req->file, req->sync.off, in io_fsync()
4287 req->sync.off = READ_ONCE(sqe->off); in io_fallocate_prep()
4300 ret = vfs_fallocate(req->file, req->sync.mode, req->sync.off, in io_fallocate()
4451 if (sqe->ioprio || sqe->rw_flags || sqe->addr || sqe->len || sqe->off || in io_remove_buffers_prep()
4545 tmp = READ_ONCE(sqe->off); in io_provide_buffers_prep()
4620 req->epoll.fd = READ_ONCE(sqe->off); in io_epoll_ctl_prep()
4659 if (sqe->ioprio || sqe->buf_index || sqe->off || sqe->splice_fd_in) in io_madvise_prep()
4699 req->fadvise.offset = READ_ONCE(sqe->off); in io_fadvise_prep()
4767 if (sqe->ioprio || sqe->off || sqe->addr || sqe->len || in io_close_prep()
4842 req->sync.off = READ_ONCE(sqe->off); in io_sfr_prep()
4856 ret = sync_file_range(req->file, req->sync.off, req->sync.len, in io_sync_file_range()
6106 upd->new_user_data = READ_ONCE(sqe->off); in io_poll_update_prep()
6124 if (sqe->ioprio || sqe->buf_index || sqe->off || sqe->addr) in io_poll_add_prep()
6309 req->timeout.off = 0; /* noseq */ in io_timeout_update()
6393 u32 off = READ_ONCE(sqe->off); in io_timeout_prep() local
6400 if (off && is_timeout_link) in io_timeout_prep()
6410 req->timeout.off = off; in io_timeout_prep()
6411 if (unlikely(off && !req->ctx->off_timeout_used)) in io_timeout_prep()
6446 u32 tail, off = req->timeout.off; in io_timeout() local
6461 req->timeout.target_seq = tail + off; in io_timeout()
6480 if (off >= nxt->timeout.target_seq - tail) in io_timeout()
6560 if (sqe->ioprio || sqe->off || sqe->len || sqe->cancel_flags || in io_async_cancel_prep()
6605 req->rsrc_update.offset = READ_ONCE(sqe->off); in io_rsrc_update_prep()
8097 unsigned int off = idx & IO_RSRC_TAG_TABLE_MASK; in io_get_tag_slot() local
8100 return &data->tags[table_idx][off]; in io_get_tag_slot()
9050 size_t off, sq_array_size; in rings_size() local
9052 off = struct_size(rings, cqes, cq_entries); in rings_size()
9053 if (off == SIZE_MAX) in rings_size()
9057 off = ALIGN(off, SMP_CACHE_BYTES); in rings_size()
9058 if (off == 0) in rings_size()
9063 *sq_offset = off; in rings_size()
9069 if (check_add_overflow(off, sq_array_size, &off)) in rings_size()
9072 return off; in rings_size()
9230 unsigned long off, start, end, ubuf; in io_sqe_buffer_register() local
9300 off = ubuf & ~PAGE_MASK; in io_sqe_buffer_register()
9305 vec_len = min_t(size_t, size, PAGE_SIZE - off); in io_sqe_buffer_register()
9308 imu->bvec[i].bv_offset = off; in io_sqe_buffer_register()
9309 off = 0; in io_sqe_buffer_register()
11280 BUILD_BUG_SQE_ELEM(8, __u64, off); in io_uring_init()