Lines Matching refs:v
84 #define iterate_all_kinds(i, n, v, I, B, K) { \ argument
88 struct bio_vec v; \
89 iterate_bvec(i, n, v, bvec, skip, (B)) \
92 struct kvec v; \
93 iterate_kvec(i, n, v, kvec, skip, (K)) \
96 struct iovec v; \
97 iterate_iovec(i, n, v, iov, skip, (I)) \
101 #define iterate_and_advance(i, n, v, I, B, K) { \ argument
105 struct bio_vec v; \
106 iterate_bvec(i, n, v, bvec, skip, (B)) \
115 struct kvec v; \
116 iterate_kvec(i, n, v, kvec, skip, (K)) \
125 struct iovec v; \
126 iterate_iovec(i, n, v, iov, skip, (I)) \
312 struct iovec v; in iov_iter_fault_in_readable() local
315 iterate_iovec(i, bytes, v, iov, skip, ({ in iov_iter_fault_in_readable()
316 err = fault_in_multipages_readable(v.iov_base, in iov_iter_fault_in_readable()
317 v.iov_len); in iov_iter_fault_in_readable()
375 iterate_and_advance(i, bytes, v, in copy_to_iter()
376 __copy_to_user(v.iov_base, (from += v.iov_len) - v.iov_len, in copy_to_iter()
377 v.iov_len), in copy_to_iter()
378 memcpy_to_page(v.bv_page, v.bv_offset, in copy_to_iter()
379 (from += v.bv_len) - v.bv_len, v.bv_len), in copy_to_iter()
380 memcpy(v.iov_base, (from += v.iov_len) - v.iov_len, v.iov_len) in copy_to_iter()
396 iterate_and_advance(i, bytes, v, in copy_from_iter()
397 __copy_from_user((to += v.iov_len) - v.iov_len, v.iov_base, in copy_from_iter()
398 v.iov_len), in copy_from_iter()
399 memcpy_from_page((to += v.bv_len) - v.bv_len, v.bv_page, in copy_from_iter()
400 v.bv_offset, v.bv_len), in copy_from_iter()
401 memcpy((to += v.iov_len) - v.iov_len, v.iov_base, v.iov_len) in copy_from_iter()
417 iterate_and_advance(i, bytes, v, in copy_from_iter_nocache()
418 __copy_from_user_nocache((to += v.iov_len) - v.iov_len, in copy_from_iter_nocache()
419 v.iov_base, v.iov_len), in copy_from_iter_nocache()
420 memcpy_from_page((to += v.bv_len) - v.bv_len, v.bv_page, in copy_from_iter_nocache()
421 v.bv_offset, v.bv_len), in copy_from_iter_nocache()
422 memcpy((to += v.iov_len) - v.iov_len, v.iov_base, v.iov_len) in copy_from_iter_nocache()
463 iterate_and_advance(i, bytes, v, in iov_iter_zero()
464 __clear_user(v.iov_base, v.iov_len), in iov_iter_zero()
465 memzero_page(v.bv_page, v.bv_offset, v.bv_len), in iov_iter_zero()
466 memset(v.iov_base, 0, v.iov_len) in iov_iter_zero()
477 iterate_all_kinds(i, bytes, v, in iov_iter_copy_from_user_atomic()
478 __copy_from_user_inatomic((p += v.iov_len) - v.iov_len, in iov_iter_copy_from_user_atomic()
479 v.iov_base, v.iov_len), in iov_iter_copy_from_user_atomic()
480 memcpy_from_page((p += v.bv_len) - v.bv_len, v.bv_page, in iov_iter_copy_from_user_atomic()
481 v.bv_offset, v.bv_len), in iov_iter_copy_from_user_atomic()
482 memcpy((p += v.iov_len) - v.iov_len, v.iov_base, v.iov_len) in iov_iter_copy_from_user_atomic()
491 iterate_and_advance(i, size, v, 0, 0, 0) in iov_iter_advance()
543 iterate_all_kinds(i, size, v, in iov_iter_alignment()
544 (res |= (unsigned long)v.iov_base | v.iov_len, 0), in iov_iter_alignment()
545 res |= v.bv_offset | v.bv_len, in iov_iter_alignment()
546 res |= (unsigned long)v.iov_base | v.iov_len in iov_iter_alignment()
562 iterate_all_kinds(i, maxsize, v, ({ in iov_iter_get_pages()
563 unsigned long addr = (unsigned long)v.iov_base; in iov_iter_get_pages()
564 size_t len = v.iov_len + (*start = addr & (PAGE_SIZE - 1)); in iov_iter_get_pages()
578 *start = v.bv_offset; in iov_iter_get_pages()
579 get_page(*pages = v.bv_page); in iov_iter_get_pages()
580 return v.bv_len; in iov_iter_get_pages()
609 iterate_all_kinds(i, maxsize, v, ({ in iov_iter_get_pages_alloc()
610 unsigned long addr = (unsigned long)v.iov_base; in iov_iter_get_pages_alloc()
611 size_t len = v.iov_len + (*start = addr & (PAGE_SIZE - 1)); in iov_iter_get_pages_alloc()
629 *start = v.bv_offset; in iov_iter_get_pages_alloc()
633 get_page(*p = v.bv_page); in iov_iter_get_pages_alloc()
634 return v.bv_len; in iov_iter_get_pages_alloc()
656 iterate_and_advance(i, bytes, v, ({ in csum_and_copy_from_iter()
658 next = csum_and_copy_from_user(v.iov_base, in csum_and_copy_from_iter()
659 (to += v.iov_len) - v.iov_len, in csum_and_copy_from_iter()
660 v.iov_len, 0, &err); in csum_and_copy_from_iter()
663 off += v.iov_len; in csum_and_copy_from_iter()
665 err ? v.iov_len : 0; in csum_and_copy_from_iter()
667 char *p = kmap_atomic(v.bv_page); in csum_and_copy_from_iter()
668 next = csum_partial_copy_nocheck(p + v.bv_offset, in csum_and_copy_from_iter()
669 (to += v.bv_len) - v.bv_len, in csum_and_copy_from_iter()
670 v.bv_len, 0); in csum_and_copy_from_iter()
673 off += v.bv_len; in csum_and_copy_from_iter()
675 next = csum_partial_copy_nocheck(v.iov_base, in csum_and_copy_from_iter()
676 (to += v.iov_len) - v.iov_len, in csum_and_copy_from_iter()
677 v.iov_len, 0); in csum_and_copy_from_iter()
679 off += v.iov_len; in csum_and_copy_from_iter()
700 iterate_and_advance(i, bytes, v, ({ in csum_and_copy_to_iter()
702 next = csum_and_copy_to_user((from += v.iov_len) - v.iov_len, in csum_and_copy_to_iter()
703 v.iov_base, in csum_and_copy_to_iter()
704 v.iov_len, 0, &err); in csum_and_copy_to_iter()
707 off += v.iov_len; in csum_and_copy_to_iter()
709 err ? v.iov_len : 0; in csum_and_copy_to_iter()
711 char *p = kmap_atomic(v.bv_page); in csum_and_copy_to_iter()
712 next = csum_partial_copy_nocheck((from += v.bv_len) - v.bv_len, in csum_and_copy_to_iter()
713 p + v.bv_offset, in csum_and_copy_to_iter()
714 v.bv_len, 0); in csum_and_copy_to_iter()
717 off += v.bv_len; in csum_and_copy_to_iter()
719 next = csum_partial_copy_nocheck((from += v.iov_len) - v.iov_len, in csum_and_copy_to_iter()
720 v.iov_base, in csum_and_copy_to_iter()
721 v.iov_len, 0); in csum_and_copy_to_iter()
723 off += v.iov_len; in csum_and_copy_to_iter()
739 iterate_all_kinds(i, size, v, ({ in iov_iter_npages()
740 unsigned long p = (unsigned long)v.iov_base; in iov_iter_npages()
741 npages += DIV_ROUND_UP(p + v.iov_len, PAGE_SIZE) in iov_iter_npages()
750 unsigned long p = (unsigned long)v.iov_base; in iov_iter_npages()
751 npages += DIV_ROUND_UP(p + v.iov_len, PAGE_SIZE) in iov_iter_npages()