• Home
  • Raw
  • Download

Lines Matching refs:sector_nr

45 static void allow_barrier(struct r1conf *conf, sector_t sector_nr);
46 static void lower_barrier(struct r1conf *conf, sector_t sector_nr);
881 static int raise_barrier(struct r1conf *conf, sector_t sector_nr) in raise_barrier() argument
883 int idx = sector_to_idx(sector_nr); in raise_barrier()
931 static void lower_barrier(struct r1conf *conf, sector_t sector_nr) in lower_barrier() argument
933 int idx = sector_to_idx(sector_nr); in lower_barrier()
1001 static void wait_read_barrier(struct r1conf *conf, sector_t sector_nr) in wait_read_barrier() argument
1003 int idx = sector_to_idx(sector_nr); in wait_read_barrier()
1034 static void wait_barrier(struct r1conf *conf, sector_t sector_nr) in wait_barrier() argument
1036 int idx = sector_to_idx(sector_nr); in wait_barrier()
1047 static void allow_barrier(struct r1conf *conf, sector_t sector_nr) in allow_barrier() argument
1049 int idx = sector_to_idx(sector_nr); in allow_barrier()
2631 static sector_t raid1_sync_request(struct mddev *mddev, sector_t sector_nr, in raid1_sync_request() argument
2646 int idx = sector_to_idx(sector_nr); in raid1_sync_request()
2654 if (sector_nr >= max_sector) { in raid1_sync_request()
2681 return max_sector - sector_nr; in raid1_sync_request()
2686 if (!md_bitmap_start_sync(mddev->bitmap, sector_nr, &sync_blocks, 1) && in raid1_sync_request()
2704 md_bitmap_cond_end_sync(mddev->bitmap, sector_nr, in raid1_sync_request()
2705 mddev_is_clustered(mddev) && (sector_nr + 2 * RESYNC_SECTORS > conf->cluster_sync_high)); in raid1_sync_request()
2708 if (raise_barrier(conf, sector_nr)) in raid1_sync_request()
2724 r1_bio->sector = sector_nr; in raid1_sync_request()
2728 good_sectors = align_to_barrier_unit_end(sector_nr, good_sectors); in raid1_sync_request()
2748 if (is_badblock(rdev, sector_nr, good_sectors, in raid1_sync_request()
2750 if (first_bad > sector_nr) in raid1_sync_request()
2751 good_sectors = first_bad - sector_nr; in raid1_sync_request()
2753 bad_sectors -= (sector_nr - first_bad); in raid1_sync_request()
2759 if (sector_nr < first_bad) { in raid1_sync_request()
2786 bio->bi_iter.bi_sector = sector_nr + rdev->data_offset; in raid1_sync_request()
2805 ok = rdev_set_badblocks(rdev, sector_nr, in raid1_sync_request()
2842 max_sector = sector_nr + min_bad; in raid1_sync_request()
2843 rv = max_sector - sector_nr; in raid1_sync_request()
2851 if (max_sector > sector_nr + good_sectors) in raid1_sync_request()
2852 max_sector = sector_nr + good_sectors; in raid1_sync_request()
2858 if (sector_nr + (len>>9) > max_sector) in raid1_sync_request()
2859 len = (max_sector - sector_nr) << 9; in raid1_sync_request()
2863 if (!md_bitmap_start_sync(mddev->bitmap, sector_nr, in raid1_sync_request()
2888 sector_nr += len>>9; in raid1_sync_request()
2895 conf->cluster_sync_high < sector_nr + nr_sectors) { in raid1_sync_request()