/fs/nfs/objlayout/ |
D | pnfs_osd_xdr_cli.c | 160 struct pnfs_osd_xdr_decode_layout_iter *iter, struct xdr_stream *xdr) in pnfs_osd_xdr_decode_layout_map() argument 164 memset(iter, 0, sizeof(*iter)); in pnfs_osd_xdr_decode_layout_map() 176 iter->total_comps = layout->olo_num_comps; in pnfs_osd_xdr_decode_layout_map() 181 struct pnfs_osd_xdr_decode_layout_iter *iter, struct xdr_stream *xdr, in pnfs_osd_xdr_decode_layout_comp() argument 184 BUG_ON(iter->decoded_comps > iter->total_comps); in pnfs_osd_xdr_decode_layout_comp() 185 if (iter->decoded_comps == iter->total_comps) in pnfs_osd_xdr_decode_layout_comp() 192 iter->decoded_comps, iter->total_comps); in pnfs_osd_xdr_decode_layout_comp() 204 iter->decoded_comps++; in pnfs_osd_xdr_decode_layout_comp()
|
/fs/nfs/ |
D | direct.c | 259 ssize_t nfs_direct_IO(struct kiocb *iocb, struct iov_iter *iter, loff_t pos) in nfs_direct_IO() argument 267 VM_BUG_ON(iov_iter_count(iter) != PAGE_SIZE); in nfs_direct_IO() 269 if (iov_iter_rw(iter) == READ) in nfs_direct_IO() 270 return nfs_file_direct_read(iocb, iter, pos); in nfs_direct_IO() 271 return nfs_file_direct_write(iocb, iter); in nfs_direct_IO() 475 struct iov_iter *iter, in nfs_direct_read_schedule_iovec() argument 490 while (iov_iter_count(iter)) { in nfs_direct_read_schedule_iovec() 496 result = iov_iter_get_pages_alloc(iter, &pagevec, in nfs_direct_read_schedule_iovec() 502 iov_iter_advance(iter, bytes); in nfs_direct_read_schedule_iovec() 570 ssize_t nfs_file_direct_read(struct kiocb *iocb, struct iov_iter *iter, in nfs_file_direct_read() argument [all …]
|
/fs/ |
D | read_write.c | 332 ssize_t vfs_iter_read(struct file *file, struct iov_iter *iter, loff_t *ppos) in vfs_iter_read() argument 343 iter->type |= READ; in vfs_iter_read() 344 ret = file->f_op->read_iter(&kiocb, iter); in vfs_iter_read() 352 ssize_t vfs_iter_write(struct file *file, struct iov_iter *iter, loff_t *ppos) in vfs_iter_write() argument 363 iter->type |= WRITE; in vfs_iter_write() 364 ret = file->f_op->write_iter(&kiocb, iter); in vfs_iter_write() 417 struct iov_iter iter; in new_sync_read() local 422 iov_iter_init(&iter, READ, &iov, 1, len); in new_sync_read() 424 ret = filp->f_op->read_iter(&kiocb, &iter); in new_sync_read() 473 struct iov_iter iter; in new_sync_write() local [all …]
|
D | direct-io.c | 98 struct iov_iter *iter; member 163 ret = iov_iter_get_pages(sdio->iter, dio->pages, LONG_MAX, DIO_PAGES, in dio_refill_pages() 185 iov_iter_advance(sdio->iter, ret); in dio_refill_pages() 1111 struct block_device *bdev, struct iov_iter *iter, in do_blockdev_direct_IO() argument 1119 size_t count = iov_iter_count(iter); in do_blockdev_direct_IO() 1125 unsigned long align = offset | iov_iter_alignment(iter); in do_blockdev_direct_IO() 1141 if (iov_iter_rw(iter) == READ && !iov_iter_count(iter)) in do_blockdev_direct_IO() 1157 if (iov_iter_rw(iter) == READ) { in do_blockdev_direct_IO() 1176 if (iov_iter_rw(iter) == READ && offset >= dio->i_size) { in do_blockdev_direct_IO() 1193 iov_iter_rw(iter) == WRITE && end > i_size_read(inode)) in do_blockdev_direct_IO() [all …]
|
D | dax.c | 108 static ssize_t dax_io(struct inode *inode, struct iov_iter *iter, in dax_io() argument 120 if (iov_iter_rw(iter) != WRITE) in dax_io() 136 iov_iter_rw(iter) == WRITE); in dax_io() 149 hole = iov_iter_rw(iter) != WRITE && !buffer_written(bh); in dax_io() 168 if (iov_iter_rw(iter) == WRITE) { in dax_io() 169 len = copy_from_iter_pmem(addr, max - pos, iter); in dax_io() 173 iter); in dax_io() 175 len = iov_iter_zero(max - pos, iter); in dax_io() 210 struct iov_iter *iter, loff_t pos, get_block_t get_block, in dax_do_io() argument 215 loff_t end = pos + iov_iter_count(iter); in dax_do_io() [all …]
|
D | aio.c | 1409 struct iov_iter *iter) in aio_setup_vectored_rw() argument 1415 len, UIO_FASTIOV, iovec, iter); in aio_setup_vectored_rw() 1418 len, UIO_FASTIOV, iovec, iter); in aio_setup_vectored_rw() 1434 struct iov_iter iter; in aio_run_iocb() local 1459 &iovec, compat, &iter); in aio_run_iocb() 1461 ret = import_single_range(rw, buf, len, iovec, &iter); in aio_run_iocb() 1466 iov_iter_count(&iter)); in aio_run_iocb() 1477 ret = iter_op(req, &iter); in aio_run_iocb()
|
D | locks.c | 2620 struct locks_iterator *iter = f->private; in locks_show() local 2625 lock_get_status(f, fl, iter->li_pos, ""); in locks_show() 2628 lock_get_status(f, bfl, iter->li_pos, " ->"); in locks_show() 2674 struct locks_iterator *iter = f->private; in locks_start() local 2676 iter->li_pos = *pos + 1; in locks_start() 2679 return seq_hlist_start_percpu(&file_lock_list, &iter->li_cpu, *pos); in locks_start() 2684 struct locks_iterator *iter = f->private; in locks_next() local 2686 ++iter->li_pos; in locks_next() 2687 return seq_hlist_next_percpu(v, &file_lock_list, &iter->li_cpu, pos); in locks_next()
|
/fs/ocfs2/dlm/ |
D | dlmmaster.c | 733 struct dlm_node_iter iter; in dlm_get_lock_resource() local 956 dlm_node_iter_init(mle->vote_map, &iter); in dlm_get_lock_resource() 957 while ((nodenum = dlm_node_iter_next(&iter)) >= 0) { in dlm_get_lock_resource() 1191 static void dlm_bitmap_diff_iter_init(struct dlm_bitmap_diff_iter *iter, in dlm_bitmap_diff_iter_init() argument 1198 iter->curnode = -1; in dlm_bitmap_diff_iter_init() 1199 iter->orig_bm = orig_bm; in dlm_bitmap_diff_iter_init() 1200 iter->cur_bm = cur_bm; in dlm_bitmap_diff_iter_init() 1203 p1 = *(iter->orig_bm + i); in dlm_bitmap_diff_iter_init() 1204 p2 = *(iter->cur_bm + i); in dlm_bitmap_diff_iter_init() 1205 iter->diff_bm[i] = (p1 & ~p2) | (p2 & ~p1); in dlm_bitmap_diff_iter_init() [all …]
|
D | dlmcommon.h | 1112 struct dlm_node_iter *iter) in dlm_node_iter_init() argument 1114 memcpy(iter->node_map, map, sizeof(iter->node_map)); in dlm_node_iter_init() 1115 iter->curnode = -1; in dlm_node_iter_init() 1118 static inline int dlm_node_iter_next(struct dlm_node_iter *iter) in dlm_node_iter_next() argument 1121 bit = find_next_bit(iter->node_map, O2NM_MAX_NODES, iter->curnode+1); in dlm_node_iter_next() 1123 iter->curnode = O2NM_MAX_NODES; in dlm_node_iter_next() 1126 iter->curnode = bit; in dlm_node_iter_next()
|
D | dlmrecovery.c | 1100 struct list_head *iter, *queue = &res->granted; in dlm_num_locks_in_lockres() local 1104 list_for_each(iter, queue) in dlm_num_locks_in_lockres() 1605 struct dlm_node_iter iter; in dlm_lockres_master_requery() local 1635 dlm_node_iter_init(dlm->domain_map, &iter); in dlm_lockres_master_requery() 1638 while ((nodenum = dlm_node_iter_next(&iter)) >= 0) { in dlm_lockres_master_requery() 1792 struct list_head *queue, *iter; in dlm_process_recovery_data() local 1837 list_for_each(iter, tmpq) { in dlm_process_recovery_data() 1838 lock = list_entry(iter, in dlm_process_recovery_data() 2664 struct dlm_node_iter iter; in dlm_send_begin_reco_message() local 2671 dlm_node_iter_init(dlm->domain_map, &iter); in dlm_send_begin_reco_message() [all …]
|
/fs/f2fs/ |
D | trace.c | 138 struct radix_tree_iter iter; in gang_lookup_pids() local 145 radix_tree_for_each_slot(slot, &pids, &iter, first_index) { in gang_lookup_pids() 146 results[ret] = iter.index; in gang_lookup_pids()
|
D | data.c | 2380 static int check_direct_IO(struct inode *inode, struct iov_iter *iter, in check_direct_IO() argument 2388 if (iov_iter_alignment(iter) & blocksize_mask) in check_direct_IO() 2394 static ssize_t f2fs_direct_IO(struct kiocb *iocb, struct iov_iter *iter, in f2fs_direct_IO() argument 2400 size_t count = iov_iter_count(iter); in f2fs_direct_IO() 2401 int rw = iov_iter_rw(iter); in f2fs_direct_IO() 2406 err = check_direct_IO(inode, iter, offset); in f2fs_direct_IO() 2414 (iov_iter_rw(iter) == READ)) { in f2fs_direct_IO() 2425 (iov_iter_rw(iter) == WRITE)) { in f2fs_direct_IO() 2449 err = blockdev_direct_IO(iocb, inode, iter, offset, get_data_block_dio); in f2fs_direct_IO() 2465 (iov_iter_rw(iter) == READ)) in f2fs_direct_IO() [all …]
|
/fs/9p/ |
D | vfs_addr.c | 250 v9fs_direct_IO(struct kiocb *iocb, struct iov_iter *iter, loff_t pos) in v9fs_direct_IO() argument 255 if (iov_iter_rw(iter) == WRITE) { in v9fs_direct_IO() 256 n = p9_client_write(file->private_data, pos, iter, &err); in v9fs_direct_IO() 264 n = p9_client_read(file->private_data, pos, iter, &err); in v9fs_direct_IO()
|
/fs/proc/ |
D | base.c | 3152 static struct tgid_iter next_tgid(struct pid_namespace *ns, struct tgid_iter iter) in next_tgid() argument 3156 if (iter.task) in next_tgid() 3157 put_task_struct(iter.task); in next_tgid() 3160 iter.task = NULL; in next_tgid() 3161 pid = find_ge_pid(iter.tgid, ns); in next_tgid() 3163 iter.tgid = pid_nr_ns(pid, ns); in next_tgid() 3164 iter.task = pid_task(pid, PIDTYPE_PID); in next_tgid() 3177 if (!iter.task || !has_group_leader_pid(iter.task)) { in next_tgid() 3178 iter.tgid += 1; in next_tgid() 3181 get_task_struct(iter.task); in next_tgid() [all …]
|
/fs/jfs/ |
D | inode.c | 335 static ssize_t jfs_direct_IO(struct kiocb *iocb, struct iov_iter *iter, in jfs_direct_IO() argument 341 size_t count = iov_iter_count(iter); in jfs_direct_IO() 344 ret = blockdev_direct_IO(iocb, inode, iter, offset, jfs_get_block); in jfs_direct_IO() 350 if (unlikely(iov_iter_rw(iter) == WRITE && ret < 0)) { in jfs_direct_IO()
|
/fs/ext4/ |
D | indirect.c | 647 ssize_t ext4_ind_direct_IO(struct kiocb *iocb, struct iov_iter *iter, in ext4_ind_direct_IO() argument 656 size_t count = iov_iter_count(iter); in ext4_ind_direct_IO() 659 if (iov_iter_rw(iter) == WRITE) { in ext4_ind_direct_IO() 681 if (iov_iter_rw(iter) == READ && ext4_should_dioread_nolock(inode)) { in ext4_ind_direct_IO() 695 ret = dax_do_io(iocb, inode, iter, offset, in ext4_ind_direct_IO() 699 inode->i_sb->s_bdev, iter, in ext4_ind_direct_IO() 706 ret = dax_do_io(iocb, inode, iter, offset, in ext4_ind_direct_IO() 709 ret = blockdev_direct_IO(iocb, inode, iter, offset, in ext4_ind_direct_IO() 712 if (unlikely(iov_iter_rw(iter) == WRITE && ret < 0)) { in ext4_ind_direct_IO()
|
/fs/sdcardfs/ |
D | file.c | 378 ssize_t sdcardfs_read_iter(struct kiocb *iocb, struct iov_iter *iter) in sdcardfs_read_iter() argument 391 err = lower_file->f_op->read_iter(iocb, iter); in sdcardfs_read_iter() 405 ssize_t sdcardfs_write_iter(struct kiocb *iocb, struct iov_iter *iter) in sdcardfs_write_iter() argument 419 err = lower_file->f_op->write_iter(iocb, iter); in sdcardfs_write_iter()
|
D | mmap.c | 70 struct iov_iter *iter, loff_t pos) in sdcardfs_direct_IO() argument
|
/fs/fuse/ |
D | file.c | 1339 ssize_t fuse_direct_io(struct fuse_io_priv *io, struct iov_iter *iter, in fuse_direct_io() argument 1350 size_t count = iov_iter_count(iter); in fuse_direct_io() 1357 req = fuse_get_req_for_background(fc, fuse_iter_npages(iter)); in fuse_direct_io() 1359 req = fuse_get_req(fc, fuse_iter_npages(iter)); in fuse_direct_io() 1371 io->should_dirty = !write && iter_is_iovec(iter); in fuse_direct_io() 1376 int err = fuse_get_user_pages(req, iter, &nbytes, write); in fuse_direct_io() 1406 fuse_iter_npages(iter)); in fuse_direct_io() 1408 req = fuse_get_req(fc, fuse_iter_npages(iter)); in fuse_direct_io() 1423 struct iov_iter *iter, in __fuse_direct_read() argument 1433 res = fuse_direct_io(io, iter, ppos, 0); in __fuse_direct_read() [all …]
|
/fs/btrfs/tests/ |
D | btrfs-tests.c | 135 struct radix_tree_iter iter; in btrfs_free_dummy_fs_info() local 140 radix_tree_for_each_slot(slot, &fs_info->buffer_radix, &iter, 0) { in btrfs_free_dummy_fs_info()
|
/fs/ocfs2/ |
D | journal.c | 2040 struct inode *iter; in ocfs2_orphan_filldir() local 2054 iter = ocfs2_iget(p->osb, ino, in ocfs2_orphan_filldir() 2056 if (IS_ERR(iter)) in ocfs2_orphan_filldir() 2061 OCFS2_I(iter)->ip_flags |= OCFS2_INODE_DIO_ORPHAN_ENTRY; in ocfs2_orphan_filldir() 2065 if (OCFS2_I(iter)->ip_next_orphan) { in ocfs2_orphan_filldir() 2066 iput(iter); in ocfs2_orphan_filldir() 2070 trace_ocfs2_orphan_filldir((unsigned long long)OCFS2_I(iter)->ip_blkno); in ocfs2_orphan_filldir() 2073 OCFS2_I(iter)->ip_next_orphan = p->head; in ocfs2_orphan_filldir() 2074 p->head = iter; in ocfs2_orphan_filldir() 2185 struct inode *iter; in ocfs2_recover_orphans() local [all …]
|
D | dlmglue.c | 2846 struct ocfs2_lock_res *iter, *ret = NULL; in ocfs2_dlm_next_res() local 2851 list_for_each_entry(iter, &start->l_debug_list, l_debug_list) { in ocfs2_dlm_next_res() 2853 if (&iter->l_debug_list == &dlm_debug->d_lockres_tracking) { in ocfs2_dlm_next_res() 2860 if (iter->l_ops != NULL) { in ocfs2_dlm_next_res() 2861 ret = iter; in ocfs2_dlm_next_res() 2872 struct ocfs2_lock_res *iter; in ocfs2_dlm_seq_start() local 2875 iter = ocfs2_dlm_next_res(&priv->p_iter_res, priv); in ocfs2_dlm_seq_start() 2876 if (iter) { in ocfs2_dlm_seq_start() 2884 priv->p_tmp_res = *iter; in ocfs2_dlm_seq_start() 2885 iter = &priv->p_tmp_res; in ocfs2_dlm_seq_start() [all …]
|
/fs/hfsplus/ |
D | inode.c | 125 static ssize_t hfsplus_direct_IO(struct kiocb *iocb, struct iov_iter *iter, in hfsplus_direct_IO() argument 131 size_t count = iov_iter_count(iter); in hfsplus_direct_IO() 134 ret = blockdev_direct_IO(iocb, inode, iter, offset, hfsplus_get_block); in hfsplus_direct_IO() 140 if (unlikely(iov_iter_rw(iter) == WRITE && ret < 0)) { in hfsplus_direct_IO()
|
/fs/hfs/ |
D | inode.c | 127 static ssize_t hfs_direct_IO(struct kiocb *iocb, struct iov_iter *iter, in hfs_direct_IO() argument 133 size_t count = iov_iter_count(iter); in hfs_direct_IO() 136 ret = blockdev_direct_IO(iocb, inode, iter, offset, hfs_get_block); in hfs_direct_IO() 142 if (unlikely(iov_iter_rw(iter) == WRITE && ret < 0)) { in hfs_direct_IO()
|
/fs/xfs/ |
D | xfs_aops.c | 1692 struct iov_iter *iter, in xfs_vm_do_dio() argument 1703 return dax_do_io(iocb, inode, iter, offset, in xfs_vm_do_dio() 1707 return __blockdev_direct_IO(iocb, inode, bdev, iter, offset, in xfs_vm_do_dio() 1714 struct iov_iter *iter, in xfs_vm_direct_IO() argument 1719 if (iov_iter_rw(iter) == WRITE) in xfs_vm_direct_IO() 1720 return xfs_vm_do_dio(inode, iocb, iter, offset, in xfs_vm_direct_IO() 1722 return xfs_vm_do_dio(inode, iocb, iter, offset, NULL, 0); in xfs_vm_direct_IO()
|