• Home
  • Raw
  • Download

Lines Matching refs:i

78 	unsigned int i, entry = -1;  in bio_find_or_create_slab()  local
82 i = 0; in bio_find_or_create_slab()
83 while (i < bio_slab_nr) { in bio_find_or_create_slab()
84 bslab = &bio_slabs[i]; in bio_find_or_create_slab()
87 entry = i; in bio_find_or_create_slab()
93 i++; in bio_find_or_create_slab()
131 unsigned int i; in bio_put_slab() local
135 for (i = 0; i < bio_slab_nr; i++) { in bio_put_slab()
136 if (bs->bio_slab == bio_slabs[i].slab) { in bio_put_slab()
137 bslab = &bio_slabs[i]; in bio_put_slab()
942 int i; in bio_alloc_pages() local
945 bio_for_each_segment_all(bv, bio, i) { in bio_alloc_pages()
1047 int ret = 0, i; in __bio_copy_iov() local
1052 bio_for_each_segment_all(bvec, bio, i) { in __bio_copy_iov()
1106 int ret = 0, i; in bio_uncopy_user() local
1118 bio_for_each_segment_all(bvec, bio, i) in bio_uncopy_user()
1149 int i, ret; in bio_copy_user_iov() local
1154 for (i = 0; i < iov_count; i++) { in bio_copy_user_iov()
1159 uaddr = (unsigned long)iov[i].iov_base; in bio_copy_user_iov()
1160 end = (uaddr + iov[i].iov_len + PAGE_SIZE - 1) >> PAGE_SHIFT; in bio_copy_user_iov()
1170 len += iov[i].iov_len; in bio_copy_user_iov()
1192 i = map_data->offset / PAGE_SIZE; in bio_copy_user_iov()
1203 if (i == map_data->nr_entries * nr_pages) { in bio_copy_user_iov()
1208 page = map_data->pages[i / nr_pages]; in bio_copy_user_iov()
1209 page += (i % nr_pages); in bio_copy_user_iov()
1211 i++; in bio_copy_user_iov()
1244 bio_for_each_segment_all(bvec, bio, i) in bio_copy_user_iov()
1284 int i, j; in __bio_map_user_iov() local
1291 for (i = 0; i < iov_count; i++) { in __bio_map_user_iov()
1292 unsigned long uaddr = (unsigned long)iov[i].iov_base; in __bio_map_user_iov()
1293 unsigned long len = iov[i].iov_len; in __bio_map_user_iov()
1323 for (i = 0; i < iov_count; i++) { in __bio_map_user_iov()
1324 unsigned long uaddr = (unsigned long)iov[i].iov_base; in __bio_map_user_iov()
1325 unsigned long len = iov[i].iov_len; in __bio_map_user_iov()
1388 for (i = 0; i < nr_pages; i++) { in __bio_map_user_iov()
1389 if(!pages[i]) in __bio_map_user_iov()
1391 page_cache_release(pages[i]); in __bio_map_user_iov()
1461 int i; in __bio_unmap_user() local
1466 bio_for_each_segment_all(bvec, bio, i) { in __bio_unmap_user()
1504 int offset, i; in __bio_map_kern() local
1512 for (i = 0; i < nr_pages; i++) { in __bio_map_kern()
1569 int i; in bio_copy_kern_endio() local
1572 bio_for_each_segment_all(bvec, bio, i) { in bio_copy_kern_endio()
1602 int i; in bio_copy_kern() local
1611 bio_for_each_segment_all(bvec, bio, i) { in bio_copy_kern()
1657 int i; in bio_set_pages_dirty() local
1659 bio_for_each_segment_all(bvec, bio, i) { in bio_set_pages_dirty()
1670 int i; in bio_release_pages() local
1672 bio_for_each_segment_all(bvec, bio, i) { in bio_release_pages()
1724 int i; in bio_check_pages_dirty() local
1726 bio_for_each_segment_all(bvec, bio, i) { in bio_check_pages_dirty()
2060 int i; in biovec_init_slabs() local
2062 for (i = 0; i < BIOVEC_NR_POOLS; i++) { in biovec_init_slabs()
2064 struct biovec_slab *bvs = bvec_slabs + i; in biovec_init_slabs()