• Home
  • Raw
  • Download

Lines Matching refs:devs

94 	int size = offsetof(struct r10bio, devs[conf->geo.raid_disks]);  in r10bio_pool_alloc()
151 r10_bio->devs[j].bio = bio; in r10buf_pool_alloc()
157 r10_bio->devs[j].repl_bio = bio; in r10buf_pool_alloc()
164 struct bio *rbio = r10_bio->devs[j].repl_bio; in r10buf_pool_alloc()
171 bio = r10_bio->devs[j].bio; in r10buf_pool_alloc()
199 if (r10_bio->devs[j].bio) in r10buf_pool_alloc()
200 bio_put(r10_bio->devs[j].bio); in r10buf_pool_alloc()
201 if (r10_bio->devs[j].repl_bio) in r10buf_pool_alloc()
202 bio_put(r10_bio->devs[j].repl_bio); in r10buf_pool_alloc()
218 struct bio *bio = r10bio->devs[j].bio; in r10buf_pool_free()
226 bio = r10bio->devs[j].repl_bio; in r10buf_pool_free()
242 struct bio **bio = & r10_bio->devs[i].bio; in put_all_bios()
246 bio = &r10_bio->devs[i].repl_bio; in put_all_bios()
319 conf->mirrors[r10_bio->devs[slot].devnum].head_position = in update_head_pos()
320 r10_bio->devs[slot].addr + (r10_bio->sectors); in update_head_pos()
333 if (r10_bio->devs[slot].bio == bio) in find_bio_disk()
335 if (r10_bio->devs[slot].repl_bio == bio) { in find_bio_disk()
347 return r10_bio->devs[slot].devnum; in find_bio_disk()
359 rdev = r10_bio->devs[slot].rdev; in raid10_end_read_request()
480 r10_bio->devs[slot].bio = NULL; in raid10_end_write_request()
512 r10_bio->devs[slot].addr, in raid10_end_write_request()
517 r10_bio->devs[slot].repl_bio = IO_MADE_GOOD; in raid10_end_write_request()
519 r10_bio->devs[slot].bio = IO_MADE_GOOD; in raid10_end_write_request()
595 r10bio->devs[slot].devnum = d; in __raid10_find_phys()
596 r10bio->devs[slot].addr = s; in __raid10_find_phys()
613 r10bio->devs[slot].devnum = d; in __raid10_find_phys()
614 r10bio->devs[slot].addr = s; in __raid10_find_phys()
750 if (r10_bio->devs[slot].bio == IO_BLOCKED) in read_balance()
752 disk = r10_bio->devs[slot].devnum; in read_balance()
755 r10_bio->devs[slot].addr + sectors > in read_balance()
769 r10_bio->devs[slot].addr + sectors > rdev->recovery_offset) in read_balance()
772 dev_sector = r10_bio->devs[slot].addr; in read_balance()
828 new_distance = r10_bio->devs[slot].addr; in read_balance()
830 new_distance = abs(r10_bio->devs[slot].addr - in read_balance()
1169 if (slot >= 0 && r10_bio->devs[slot].rdev) { in raid10_read_request()
1185 disk = r10_bio->devs[slot].devnum; in raid10_read_request()
1192 err_rdev = r10_bio->devs[slot].rdev; in raid10_read_request()
1232 r10_bio->devs[slot].bio = read_bio; in raid10_read_request()
1233 r10_bio->devs[slot].rdev = rdev; in raid10_read_request()
1235 read_bio->bi_iter.bi_sector = r10_bio->devs[slot].addr + in raid10_read_request()
1264 int devnum = r10_bio->devs[n_copy].devnum; in raid10_write_one_disk()
1279 r10_bio->devs[n_copy].repl_bio = mbio; in raid10_write_one_disk()
1281 r10_bio->devs[n_copy].bio = mbio; in raid10_write_one_disk()
1283 mbio->bi_iter.bi_sector = (r10_bio->devs[n_copy].addr + in raid10_write_one_disk()
1364 sector_t dev_sector = r10_bio->devs[i].addr; in wait_blocked_dev()
1476 int d = r10_bio->devs[i].devnum; in raid10_write_request()
1485 r10_bio->devs[i].bio = NULL; in raid10_write_request()
1486 r10_bio->devs[i].repl_bio = NULL; in raid10_write_request()
1494 sector_t dev_sector = r10_bio->devs[i].addr; in raid10_write_request()
1525 r10_bio->devs[i].bio = bio; in raid10_write_request()
1529 r10_bio->devs[i].repl_bio = bio; in raid10_write_request()
1555 if (r10_bio->devs[i].bio) in raid10_write_request()
1557 if (r10_bio->devs[i].repl_bio) in raid10_write_request()
1578 memset(r10_bio->devs, 0, sizeof(r10_bio->devs[0]) * in __make_request()
1761 memset(r10_bio->devs, 0, sizeof(r10_bio->devs[0]) * geo->raid_disks); in raid10_handle_discard()
1789 r10_bio->devs[disk].bio = NULL; in raid10_handle_discard()
1790 r10_bio->devs[disk].repl_bio = NULL; in raid10_handle_discard()
1800 r10_bio->devs[disk].bio = bio; in raid10_handle_discard()
1804 r10_bio->devs[disk].repl_bio = bio; in raid10_handle_discard()
1847 if (r10_bio->devs[disk].bio) { in raid10_handle_discard()
1852 r10_bio->devs[disk].bio = mbio; in raid10_handle_discard()
1853 r10_bio->devs[disk].devnum = disk; in raid10_handle_discard()
1860 if (r10_bio->devs[disk].repl_bio) { in raid10_handle_discard()
1865 r10_bio->devs[disk].repl_bio = rbio; in raid10_handle_discard()
1866 r10_bio->devs[disk].devnum = disk; in raid10_handle_discard()
2370 r10_bio->devs[slot].addr, in end_sync_write()
2408 if (!r10_bio->devs[i].bio->bi_status) in sync_request_write()
2415 fbio = r10_bio->devs[i].bio; in sync_request_write()
2427 tbio = r10_bio->devs[i].bio; in sync_request_write()
2435 d = r10_bio->devs[i].devnum; in sync_request_write()
2437 if (!r10_bio->devs[i].bio->bi_status) { in sync_request_write()
2476 tbio->bi_iter.bi_sector = r10_bio->devs[i].addr; in sync_request_write()
2499 tbio = r10_bio->devs[i].repl_bio; in sync_request_write()
2502 if (r10_bio->devs[i].bio->bi_end_io != end_sync_write in sync_request_write()
2503 && r10_bio->devs[i].bio != fbio) in sync_request_write()
2505 d = r10_bio->devs[i].devnum; in sync_request_write()
2540 struct bio *bio = r10_bio->devs[0].bio; in fix_recovery_read_error()
2544 int dr = r10_bio->devs[0].devnum; in fix_recovery_read_error()
2545 int dw = r10_bio->devs[1].devnum; in fix_recovery_read_error()
2558 addr = r10_bio->devs[0].addr + sect, in fix_recovery_read_error()
2566 addr = r10_bio->devs[1].addr + sect; in fix_recovery_read_error()
2590 addr = r10_bio->devs[1].addr + sect; in fix_recovery_read_error()
2616 struct bio *wbio = r10_bio->devs[1].bio; in recovery_request_write()
2617 struct bio *wbio2 = r10_bio->devs[1].repl_bio; in recovery_request_write()
2639 d = r10_bio->devs[1].devnum; in recovery_request_write()
2727 int d = r10_bio->devs[r10_bio->read_slot].devnum; in fix_read_error()
2751 r10_bio->devs[r10_bio->read_slot].bio = IO_BLOCKED; in fix_read_error()
2769 d = r10_bio->devs[sl].devnum; in fix_read_error()
2774 is_badblock(rdev, r10_bio->devs[sl].addr + sect, s, in fix_read_error()
2779 r10_bio->devs[sl].addr + in fix_read_error()
2800 int dn = r10_bio->devs[r10_bio->read_slot].devnum; in fix_read_error()
2805 r10_bio->devs[r10_bio->read_slot].addr in fix_read_error()
2809 r10_bio->devs[r10_bio->read_slot].bio in fix_read_error()
2824 d = r10_bio->devs[sl].devnum; in fix_read_error()
2834 r10_bio->devs[sl].addr + in fix_read_error()
2860 d = r10_bio->devs[sl].devnum; in fix_read_error()
2870 r10_bio->devs[sl].addr + in fix_read_error()
2911 struct md_rdev *rdev = conf->mirrors[r10_bio->devs[i].devnum].rdev; in narrow_write_error()
2947 wsector = r10_bio->devs[i].addr + (sector - r10_bio->sector); in narrow_write_error()
2972 struct md_rdev *rdev = r10_bio->devs[slot].rdev; in handle_read_error()
2982 bio = r10_bio->devs[slot].bio; in handle_read_error()
2984 r10_bio->devs[slot].bio = NULL; in handle_read_error()
2987 r10_bio->devs[slot].bio = IO_BLOCKED; in handle_read_error()
3019 int dev = r10_bio->devs[m].devnum; in handle_write_completed()
3021 if (r10_bio->devs[m].bio == NULL || in handle_write_completed()
3022 r10_bio->devs[m].bio->bi_end_io == NULL) in handle_write_completed()
3024 if (!r10_bio->devs[m].bio->bi_status) { in handle_write_completed()
3027 r10_bio->devs[m].addr, in handle_write_completed()
3032 r10_bio->devs[m].addr, in handle_write_completed()
3037 if (r10_bio->devs[m].repl_bio == NULL || in handle_write_completed()
3038 r10_bio->devs[m].repl_bio->bi_end_io == NULL) in handle_write_completed()
3041 if (!r10_bio->devs[m].repl_bio->bi_status) { in handle_write_completed()
3044 r10_bio->devs[m].addr, in handle_write_completed()
3049 r10_bio->devs[m].addr, in handle_write_completed()
3058 int dev = r10_bio->devs[m].devnum; in handle_write_completed()
3059 struct bio *bio = r10_bio->devs[m].bio; in handle_write_completed()
3064 r10_bio->devs[m].addr, in handle_write_completed()
3076 bio = r10_bio->devs[m].repl_bio; in handle_write_completed()
3081 r10_bio->devs[m].addr, in handle_write_completed()
3213 bio = r10bio->devs[i].bio; in raid10_alloc_init_r10buf()
3217 bio = r10bio->devs[i].repl_bio; in raid10_alloc_init_r10buf()
3531 int d = r10_bio->devs[j].devnum; in raid10_sync_request()
3542 sector = r10_bio->devs[j].addr; in raid10_sync_request()
3556 bio = r10_bio->devs[0].bio; in raid10_sync_request()
3563 from_addr = r10_bio->devs[j].addr; in raid10_sync_request()
3571 if (r10_bio->devs[k].devnum == i) in raid10_sync_request()
3574 to_addr = r10_bio->devs[k].addr; in raid10_sync_request()
3575 r10_bio->devs[0].devnum = d; in raid10_sync_request()
3576 r10_bio->devs[0].addr = from_addr; in raid10_sync_request()
3577 r10_bio->devs[1].devnum = i; in raid10_sync_request()
3578 r10_bio->devs[1].addr = to_addr; in raid10_sync_request()
3581 bio = r10_bio->devs[1].bio; in raid10_sync_request()
3591 r10_bio->devs[1].bio->bi_end_io = NULL; in raid10_sync_request()
3594 bio = r10_bio->devs[1].repl_bio; in raid10_sync_request()
3623 if (r10_bio->devs[k].devnum == i) in raid10_sync_request()
3629 r10_bio->devs[k].addr, in raid10_sync_request()
3635 r10_bio->devs[k].addr, in raid10_sync_request()
3659 if (r10_bio->devs[0].bio->bi_opf & MD_FAILFAST) { in raid10_sync_request()
3666 int d = r10_bio->devs[j].devnum; in raid10_sync_request()
3673 r10_bio->devs[0].bio->bi_opf in raid10_sync_request()
3726 int d = r10_bio->devs[i].devnum; in raid10_sync_request()
3731 if (r10_bio->devs[i].repl_bio) in raid10_sync_request()
3732 r10_bio->devs[i].repl_bio->bi_end_io = NULL; in raid10_sync_request()
3734 bio = r10_bio->devs[i].bio; in raid10_sync_request()
3742 sector = r10_bio->devs[i].addr; in raid10_sync_request()
3775 bio = r10_bio->devs[i].repl_bio; in raid10_sync_request()
3778 sector = r10_bio->devs[i].addr; in raid10_sync_request()
3793 int d = r10_bio->devs[i].devnum; in raid10_sync_request()
3794 if (r10_bio->devs[i].bio->bi_end_io) in raid10_sync_request()
3797 if (r10_bio->devs[i].repl_bio && in raid10_sync_request()
3798 r10_bio->devs[i].repl_bio->bi_end_io) in raid10_sync_request()
4390 static void *raid10_takeover_raid0(struct mddev *mddev, sector_t size, int devs) in raid10_takeover_raid0() argument
4400 sector_div(size, devs); in raid10_takeover_raid0()
4943 read_bio->bi_iter.bi_sector = (r10_bio->devs[r10_bio->read_slot].addr in reshape_request()
4949 r10_bio->read_slot = r10_bio->devs[r10_bio->read_slot].devnum; in reshape_request()
4986 int d = r10_bio->devs[s/2].devnum; in reshape_request()
4990 b = r10_bio->devs[s/2].repl_bio; in reshape_request()
4993 b = r10_bio->devs[s/2].bio; in reshape_request()
4999 b->bi_iter.bi_sector = r10_bio->devs[s/2].addr + in reshape_request()
5010 pages = get_resync_pages(r10_bio->devs[0].bio)->pages; in reshape_request()
5077 int d = r10_bio->devs[s/2].devnum; in reshape_request_write()
5082 b = r10_bio->devs[s/2].repl_bio; in reshape_request_write()
5085 b = r10_bio->devs[s/2].bio; in reshape_request_write()
5143 r10b = kmalloc(struct_size(r10b, devs, conf->copies), GFP_NOIO); in handle_reshape_read_error()
5150 pages = get_resync_pages(r10_bio->devs[0].bio)->pages; in handle_reshape_read_error()
5165 int d = r10b->devs[slot].devnum; in handle_reshape_read_error()
5173 addr = r10b->devs[slot].addr + idx * PAGE_SIZE; in handle_reshape_read_error()