Lines Matching refs:sectors
88 struct scrub_sector *sectors[SCRUB_SECTORS_PER_BIO]; member
100 struct scrub_sector *sectors[SCRUB_MAX_SECTORS_PER_BLOCK]; member
321 ASSERT(sblock->sectors[sblock->sector_count] == NULL); in alloc_scrub_sector()
327 sblock->sectors[sblock->sector_count] = ssector; in alloc_scrub_sector()
715 scrub_block_put(sbio->sectors[i]->sblock); in scrub_free_ctx()
909 if (sblock->sectors[0]->flags & BTRFS_EXTENT_FLAG_SUPER) { in scrub_print_warning()
1010 if (sblock_to_check->sectors[0]->flags & BTRFS_EXTENT_FLAG_SUPER) { in scrub_handle_errored_block()
1026 is_metadata = !(sblock_to_check->sectors[0]->flags & in scrub_handle_errored_block()
1028 have_csum = sblock_to_check->sectors[0]->have_csum; in scrub_handle_errored_block()
1207 if (!scrub_is_page_on_raid56(sblock_bad->sectors[0])) { in scrub_handle_errored_block()
1215 struct scrub_recover *r = sblock_bad->sectors[0]->recover; in scrub_handle_errored_block()
1276 struct scrub_sector *sector_bad = sblock_bad->sectors[sector_num]; in scrub_handle_errored_block()
1283 if (scrub_is_page_on_raid56(sblock_bad->sectors[0])) { in scrub_handle_errored_block()
1299 sectors[sector_num]->io_error) { in scrub_handle_errored_block()
1389 recover = sblock->sectors[sector_index]->recover; in scrub_handle_errored_block()
1392 sblock->sectors[sector_index]->recover = NULL; in scrub_handle_errored_block()
1451 u64 generation = original_sblock->sectors[0]->generation; in scrub_setup_recheck_block()
1452 u64 flags = original_sblock->sectors[0]->flags; in scrub_setup_recheck_block()
1453 u64 have_csum = original_sblock->sectors[0]->have_csum; in scrub_setup_recheck_block()
1519 original_sblock->sectors[0]->csum, in scrub_setup_recheck_block()
1580 struct scrub_sector *first_sector = sblock->sectors[0]; in scrub_recheck_block_on_raid56()
1592 struct scrub_sector *sector = sblock->sectors[i]; in scrub_recheck_block_on_raid56()
1609 sblock->sectors[i]->io_error = 1; in scrub_recheck_block_on_raid56()
1630 if (!retry_failed_mirror && scrub_is_page_on_raid56(sblock->sectors[0])) in scrub_recheck_block()
1634 struct scrub_sector *sector = sblock->sectors[i]; in scrub_recheck_block()
1677 if (sblock->sectors[0]->flags & BTRFS_EXTENT_FLAG_DATA) in scrub_recheck_block_checksum()
1705 struct scrub_sector *sector_bad = sblock_bad->sectors[sector_num]; in scrub_repair_sector_from_good_copy()
1706 struct scrub_sector *sector_good = sblock_good->sectors[sector_num]; in scrub_repair_sector_from_good_copy()
1766 struct scrub_sector *sector = sblock->sectors[sector_num]; in scrub_write_sector_to_dev_replace()
1861 sbio->sectors[sbio->sector_count] = sector; in scrub_add_sector_to_wr_bio()
1924 struct scrub_sector *sector = sbio->sectors[i]; in scrub_wr_bio_end_io_worker()
1936 scrub_block_put(sbio->sectors[i]->sblock); in scrub_wr_bio_end_io_worker()
1937 scrub_sector_put(sbio->sectors[i]); in scrub_wr_bio_end_io_worker()
1963 flags = sblock->sectors[0]->flags; in scrub_checksum()
1989 sector = sblock->sectors[0]; in scrub_checksum_data()
2029 sector = sblock->sectors[0]; in scrub_checksum_tree_block()
2073 kaddr = scrub_sector_get_kaddr(sblock->sectors[i]); in scrub_checksum_tree_block()
2115 sector = sblock->sectors[0]; in scrub_checksum_super()
2148 scrub_sector_put(sblock->sectors[i]); in scrub_block_put()
2292 sbio->sectors[sbio->sector_count] = sector; in scrub_add_sector_to_rd_bio()
2408 struct scrub_sector *sector = sblock->sectors[i]; in scrub_missing_raid56_pages()
2490 struct scrub_sector *sector = sblock->sectors[index]; in scrub_sectors()
2529 struct scrub_sector *sector = sbio->sectors[i]; in scrub_bio_end_io_worker()
2538 struct scrub_sector *sector = sbio->sectors[i]; in scrub_bio_end_io_worker()
2622 sblock->sectors[sblock->sector_count - 1]->offset + in scrub_block_complete()
2798 sblock->sectors[index] = sector; in scrub_sectors_for_parity()
2819 struct scrub_sector *sector = sblock->sectors[index]; in scrub_sectors_for_parity()