/fs/btrfs/ |
D | locking.h | 28 static inline void btrfs_tree_unlock_rw(struct extent_buffer *eb, int rw) in btrfs_tree_unlock_rw() argument 30 if (rw == BTRFS_WRITE_LOCK || rw == BTRFS_WRITE_LOCK_BLOCKING) in btrfs_tree_unlock_rw() 32 else if (rw == BTRFS_READ_LOCK_BLOCKING) in btrfs_tree_unlock_rw() 34 else if (rw == BTRFS_READ_LOCK) in btrfs_tree_unlock_rw()
|
D | relocation.c | 4349 int rw = 0; in btrfs_relocate_block_group() local 4375 rw = 1; in btrfs_relocate_block_group() 4451 if (err && rw) in btrfs_relocate_block_group()
|
/fs/ |
D | io_uring.c | 307 struct kiocb rw; member 490 int rw = 0; in io_queue_async_work() local 496 rw = !(req->rw.ki_flags & IOCB_DIRECT); in io_queue_async_work() 501 queue_work(ctx->sqo_wq[rw], &req->work); in io_queue_async_work() 811 struct kiocb *kiocb = &req->rw; in io_do_iopoll() 960 struct io_kiocb *req = container_of(kiocb, struct io_kiocb, rw); in io_complete_rw() 973 struct io_kiocb *req = container_of(kiocb, struct io_kiocb, rw); in io_complete_rw_iopoll() 1007 if (list_req->rw.ki_filp != req->rw.ki_filp) in io_iopoll_req_issued() 1083 struct kiocb *kiocb = &req->rw; in io_prep_rw() 1165 static int io_import_fixed(struct io_ring_ctx *ctx, int rw, in io_import_fixed() argument [all …]
|
D | aio.c | 200 struct kiocb rw; member 563 struct aio_kiocb *req = container_of(iocb, struct aio_kiocb, rw); in kiocb_set_cancel_fn() 623 req->ki_cancel(&req->rw); in free_ioctx_users() 1423 struct aio_kiocb *iocb = container_of(kiocb, struct aio_kiocb, rw); in aio_complete_rw() 1480 static ssize_t aio_setup_rw(int rw, const struct iocb *iocb, in aio_setup_rw() argument 1488 ssize_t ret = import_single_range(rw, buf, len, *iovec, iter); in aio_setup_rw() 1494 return compat_import_iovec(rw, buf, len, UIO_FASTIOV, iovec, in aio_setup_rw() 1497 return import_iovec(rw, buf, len, UIO_FASTIOV, iovec, iter); in aio_setup_rw() 1648 struct aio_kiocb *aiocb = container_of(iocb, struct aio_kiocb, rw); in aio_poll_cancel() 1813 return aio_read(&req->rw, iocb, false, compat); in __io_submit_one() [all …]
|
/fs/nfs/blocklayout/ |
D | blocklayout.h | 182 int ext_tree_remove(struct pnfs_block_layout *bl, bool rw, sector_t start, 187 struct pnfs_block_extent *ret, bool rw);
|
D | extent_tree.c | 346 struct pnfs_block_extent *ret, bool rw) in ext_tree_lookup() argument 351 if (!rw) in ext_tree_lookup() 360 int ext_tree_remove(struct pnfs_block_layout *bl, bool rw, in ext_tree_remove() argument 368 if (rw) { in ext_tree_remove()
|
D | blocklayout.c | 146 do_add_page_to_bio(struct bio *bio, int npg, int rw, sector_t isect, in do_add_page_to_bio() argument 156 npg, rw, (unsigned long long)isect, offset, *len); in do_add_page_to_bio() 183 bio_set_op_attrs(bio, rw, 0); in do_add_page_to_bio() 192 static void bl_mark_devices_unavailable(struct nfs_pgio_header *header, bool rw) in bl_mark_devices_unavailable() argument 203 if (!ext_tree_lookup(bl, isect, &be, rw)) in bl_mark_devices_unavailable()
|
/fs/crypto/ |
D | crypto.c | 90 int fscrypt_crypt_block(const struct inode *inode, fscrypt_direction_t rw, in fscrypt_crypt_block() argument 123 if (rw == FS_DECRYPT) in fscrypt_crypt_block() 130 (rw == FS_DECRYPT ? "De" : "En"), lblk_num, res); in fscrypt_crypt_block()
|
D | fscrypt_private.h | 255 fscrypt_direction_t rw, u64 lblk_num,
|
/fs/ocfs2/ |
D | refcounttree.h | 29 int ocfs2_lock_refcount_tree(struct ocfs2_super *osb, u64 ref_blkno, int rw, 34 int rw);
|
D | refcounttree.c | 416 struct ocfs2_refcount_tree *tree, int rw) in __ocfs2_lock_refcount_tree() argument 420 ret = ocfs2_refcount_lock(tree, rw); in __ocfs2_lock_refcount_tree() 426 if (rw) in __ocfs2_lock_refcount_tree() 444 u64 ref_blkno, int rw, in ocfs2_lock_refcount_tree() argument 462 ret = __ocfs2_lock_refcount_tree(osb, tree, rw); in ocfs2_lock_refcount_tree() 473 ocfs2_unlock_refcount_tree(osb, tree, rw); in ocfs2_lock_refcount_tree() 494 ocfs2_unlock_refcount_tree(osb, tree, rw); in ocfs2_lock_refcount_tree() 517 struct ocfs2_refcount_tree *tree, int rw) in ocfs2_unlock_refcount_tree() argument 519 if (rw) in ocfs2_unlock_refcount_tree() 524 ocfs2_refcount_unlock(tree, rw); in ocfs2_unlock_refcount_tree()
|
/fs/f2fs/ |
D | data.c | 2998 int rw = iov_iter_rw(iter); in f2fs_direct_IO() local 3013 trace_f2fs_direct_IO_enter(inode, offset, count, rw); in f2fs_direct_IO() 3016 (rw == READ)) { in f2fs_direct_IO() 3027 (rw == WRITE)) { in f2fs_direct_IO() 3038 if (rw == WRITE && whint_mode == WHINT_MODE_OFF) in f2fs_direct_IO() 3042 if (!down_read_trylock(&fi->i_gc_rwsem[rw])) { in f2fs_direct_IO() 3048 up_read(&fi->i_gc_rwsem[rw]); in f2fs_direct_IO() 3054 down_read(&fi->i_gc_rwsem[rw]); in f2fs_direct_IO() 3060 iter, rw == WRITE ? get_data_block_dio_write : in f2fs_direct_IO() 3067 up_read(&fi->i_gc_rwsem[rw]); in f2fs_direct_IO() [all …]
|
D | f2fs.h | 1080 #define is_read_io(rw) ((rw) == READ) argument 3725 int rw = iov_iter_rw(iter); in allow_outplace_dio() local 3727 return (test_opt(sbi, LFS) && (rw == WRITE) && in allow_outplace_dio() 3735 int rw = iov_iter_rw(iter); in f2fs_force_buffered_io() local 3753 if (test_opt(sbi, LFS) && (rw == WRITE)) { in f2fs_force_buffered_io()
|
/fs/cifs/ |
D | misc.c | 839 setup_aio_ctx_iter(struct cifs_aio_ctx *ctx, struct iov_iter *iter, int rw) in setup_aio_ctx_iter() argument 924 iov_iter_bvec(&ctx->iter, rw, ctx->bv, npages, ctx->len); in setup_aio_ctx_iter()
|
D | cifsproto.h | 578 int setup_aio_ctx_iter(struct cifs_aio_ctx *ctx, struct iov_iter *iter, int rw);
|
/fs/affs/ |
D | Changes | 225 flush the super block, and rw remounts didn't
|
/fs/ext4/ |
D | inode.c | 3894 int rw = iov_iter_rw(iter); in ext4_direct_IO() local 3916 (rw == READ)) { in ext4_direct_IO() 3927 (rw == WRITE)) { in ext4_direct_IO() 3945 (rw == READ)) in ext4_direct_IO() 3948 (rw == WRITE)) in ext4_direct_IO()
|