/fs/jffs2/ |
D | erase.c | 45 jeb->offset, jeb->offset, jeb->offset + c->sector_size); in jffs2_erase_block() 52 c->erasing_size -= c->sector_size; in jffs2_erase_block() 53 c->dirty_size += c->sector_size; in jffs2_erase_block() 54 jeb->dirty_size = c->sector_size; in jffs2_erase_block() 63 instr->len = c->sector_size; in jffs2_erase_block() 83 c->erasing_size -= c->sector_size; in jffs2_erase_block() 84 c->dirty_size += c->sector_size; in jffs2_erase_block() 85 jeb->dirty_size = c->sector_size; in jffs2_erase_block() 131 c->erasing_size += c->sector_size; in jffs2_erase_pending_blocks() 186 c->erasing_size -= c->sector_size; in jffs2_erase_failed() [all …]
|
D | build.c | 324 size += c->sector_size - 1; /* ... and round up */ in jffs2_calc_trigger_levels() 326 c->resv_blocks_write = c->resv_blocks_deletion + (size / c->sector_size); in jffs2_calc_trigger_levels() 350 c->nospc_dirty_size = c->sector_size + (c->flash_size / 100); in jffs2_calc_trigger_levels() 353 c->flash_size / 1024, c->sector_size / 1024, c->nr_blocks); in jffs2_calc_trigger_levels() 355 c->resv_blocks_deletion, c->resv_blocks_deletion*c->sector_size/1024); in jffs2_calc_trigger_levels() 357 c->resv_blocks_write, c->resv_blocks_write*c->sector_size/1024); in jffs2_calc_trigger_levels() 359 c->resv_blocks_gctrigger, c->resv_blocks_gctrigger*c->sector_size/1024); in jffs2_calc_trigger_levels() 361 c->resv_blocks_gcmerge, c->resv_blocks_gcmerge*c->sector_size/1024); in jffs2_calc_trigger_levels() 363 c->resv_blocks_gcbad, c->resv_blocks_gcbad*c->sector_size/1024); in jffs2_calc_trigger_levels() 377 c->nr_blocks = c->flash_size / c->sector_size; in jffs2_do_mount_fs() [all …]
|
D | scan.c | 62 static inline uint32_t EMPTY_SCAN_SIZE(uint32_t sector_size) { in EMPTY_SCAN_SIZE() argument 63 if (sector_size < DEFAULT_EMPTY_SCAN_SIZE) in EMPTY_SCAN_SIZE() 64 return sector_size; in EMPTY_SCAN_SIZE() 117 try_size = c->sector_size; in jffs2_scan_medium() 232 c->bad_size += c->sector_size; in jffs2_scan_medium() 233 c->free_size -= c->sector_size; in jffs2_scan_medium() 314 else if (!ISDIRTY(c->sector_size - (jeb->used_size + jeb->unchecked_size))) { in jffs2_scan_classify_jeb() 489 sm = (void *)buf + c->sector_size - sizeof(*sm); in jffs2_scan_eraseblock() 492 sumlen = c->sector_size - je32_to_cpu(sm->offset); in jffs2_scan_eraseblock() 503 jeb->offset + c->sector_size - buf_len, in jffs2_scan_eraseblock() [all …]
|
D | nodemgmt.c | 30 c->erasing_size - c->resv_blocks_write * c->sector_size in jffs2_rp_can_write() 115 …dirty = c->dirty_size + c->erasing_size - c->nr_erasing_blocks * c->sector_size + c->unchecked_siz… in jffs2_reserve_space() 124 c->sector_size); in jffs2_reserve_space() 141 if ( (avail / c->sector_size) <= blocksneeded) { in jffs2_reserve_space() 149 avail, blocksneeded * c->sector_size); in jffs2_reserve_space() 343 if (!(c->wbuf_ofs % c->sector_size) && !c->wbuf_len) in jffs2_find_nextblock() 438 (jeb->offset + c->sector_size - waste) | REF_OBSOLETE, in jffs2_do_reserve_space() 459 if (jeb->free_size != c->sector_size - c->cleanmarker_size) { in jffs2_do_reserve_space() 484 *len, jeb->offset + (c->sector_size - jeb->free_size)); in jffs2_do_reserve_space() 507 jeb = &c->blocks[ofs / c->sector_size]; in jffs2_add_physical_node_ref() [all …]
|
D | debug.c | 33 jeb->unchecked_size != c->sector_size)) { in __jffs2_dbg_acct_sanity_check_nolock() 37 jeb->wasted_size, jeb->unchecked_size, c->sector_size); in __jffs2_dbg_acct_sanity_check_nolock() 251 erasing += c->sector_size; in __jffs2_dbg_superblock_counts() 255 erasing += c->sector_size; in __jffs2_dbg_superblock_counts() 259 erasing += c->sector_size; in __jffs2_dbg_superblock_counts() 263 bad += c->sector_size; in __jffs2_dbg_superblock_counts() 322 ref_offset(ref2) > jeb->offset + c->sector_size) { in __jffs2_dbg_acct_paranoia_check_nolock() 365 && my_used_size + my_unchecked_size + my_dirty_size != c->sector_size) { in __jffs2_dbg_acct_paranoia_check_nolock() 368 c->sector_size); in __jffs2_dbg_acct_paranoia_check_nolock() 479 printk(JFFS2_DBG "sector_size: %#08x\n", c->sector_size); in __jffs2_dbg_dump_block_lists_nolock() [all …]
|
D | wbuf.c | 168 (jeb->offset+c->sector_size-oldfree) | REF_OBSOLETE, in jffs2_block_refile() 285 jeb = &c->blocks[c->wbuf_ofs / c->sector_size]; in jffs2_wbuf_recover() 460 new_jeb = &c->blocks[ofs / c->sector_size]; in jffs2_wbuf_recover() 599 wbuf_jeb = &c->blocks[c->wbuf_ofs / c->sector_size]; in __jffs2_flush_wbuf() 926 jeb = &c->blocks[outvec_to / c->sector_size]; in jffs2_flash_writev() 1253 c->sector_size = 8 * c->mtd->erasesize; in jffs2_dataflash_setup() 1255 while (c->sector_size < 8192) { in jffs2_dataflash_setup() 1256 c->sector_size *= 2; in jffs2_dataflash_setup() 1262 if ((c->flash_size % c->sector_size) != 0) { in jffs2_dataflash_setup() 1263 c->flash_size = (c->flash_size / c->sector_size) * c->sector_size; in jffs2_dataflash_setup() [all …]
|
D | fs.c | 225 if (avail > c->sector_size * c->resv_blocks_write) in jffs2_statfs() 226 avail -= c->sector_size * c->resv_blocks_write; in jffs2_statfs() 538 c->sector_size = c->mtd->erasesize; in jffs2_do_fill_super() 539 blocks = c->flash_size / c->sector_size; in jffs2_do_fill_super() 544 if ((c->sector_size * blocks) != c->flash_size) { in jffs2_do_fill_super() 545 c->flash_size = c->sector_size * blocks; in jffs2_do_fill_super() 550 if (c->flash_size < 5*c->sector_size) { in jffs2_do_fill_super() 552 c->flash_size / c->sector_size); in jffs2_do_fill_super()
|
D | summary.c | 28 uint32_t sum_size = min_t(uint32_t, c->sector_size, MAX_SUMMARY_SIZE); in jffs2_sum_init() 260 jeb = &c->blocks[ofs / c->sector_size]; in jffs2_sum_add_kvec() 376 if ((ofs & ~3) > c->sector_size - jeb->free_size) { in sum_link_node_ref() 378 jffs2_scan_dirty_space(c, jeb, (ofs & ~3) - (c->sector_size - jeb->free_size)); in sum_link_node_ref() 557 c->free_size += c->sector_size - jeb->free_size; in jffs2_sum_process_sum_data() 561 jeb->free_size = c->sector_size; in jffs2_sum_process_sum_data() 580 ofs = c->sector_size - sumsize; in jffs2_sum_scan_sumnode() 803 sm->offset = cpu_to_je32(c->sector_size - jeb->free_size); in jffs2_sum_write_data() 814 sum_ofs = jeb->offset + c->sector_size - jeb->free_size; in jffs2_sum_write_data()
|
D | nodelist.c | 615 } else if (unlikely(ref_offset(ref) != jeb->offset + c->sector_size - jeb->free_size)) { in jffs2_link_node_ref() 682 uint32_t ofs = jeb->offset + c->sector_size - jeb->free_size; in jffs2_scan_dirty_space() 703 jeb = &c->blocks[ref->flash_offset / c->sector_size]; in __ref_totlen() 713 ref_end = jeb->offset + c->sector_size - jeb->free_size; in __ref_totlen() 728 jeb = &c->blocks[ref->flash_offset / c->sector_size]; in __jffs2_ref_totlen()
|
D | nodelist.h | 206 #define write_ofs(c) ((c)->nextblock->offset + (c)->sector_size - (c)->nextblock->free_size) 298 return ((c->flash_size / c->sector_size) * sizeof (struct jffs2_eraseblock)) > (128 * 1024); in jffs2_blocks_use_vmalloc() 309 #define VERYDIRTY(c, size) ((size) >= ((c)->sector_size / 2))
|
D | os-linux.h | 65 #define SECTOR_ADDR(x) ( (((unsigned long)(x) / c->sector_size) * c->sector_size) )
|
D | jffs2_fs_sb.h | 73 uint32_t sector_size; member
|
D | write.c | 131 struct jffs2_eraseblock *jeb = &c->blocks[flash_ofs / c->sector_size]; in jffs2_write_dnode() 283 struct jffs2_eraseblock *jeb = &c->blocks[flash_ofs / c->sector_size]; in jffs2_write_dirent()
|
D | gc.c | 714 struct jffs2_eraseblock *jeb = &c->blocks[phys_ofs / c->sector_size]; in jffs2_garbage_collect_pristine() 1230 jeb = &c->blocks[raw->flash_offset / c->sector_size]; in jffs2_garbage_collect_dnode() 1286 jeb = &c->blocks[raw->flash_offset / c->sector_size]; in jffs2_garbage_collect_dnode()
|
D | readinode.c | 114 jeb = &c->blocks[ref->flash_offset / c->sector_size]; in check_node_data() 617 jeb = &c->blocks[ref->flash_offset / c->sector_size]; in read_direntry() 803 jeb = &c->blocks[ref->flash_offset / c->sector_size]; in read_dnode()
|
D | xattr.c | 185 jeb = &c->blocks[ref_offset(raw) / c->sector_size]; in do_verify_xattr_datum() 491 jeb = &c->blocks[ref_offset(raw) / c->sector_size]; in verify_xattr_ref() 1307 jeb = &c->blocks[ref_offset(raw) / c->sector_size]; in jffs2_verify_xattr()
|
/fs/ntfs/ |
D | super.c | 758 vol->sector_size = le16_to_cpu(b->bpb.bytes_per_sector); in parse_ntfs_boot_sector() 759 vol->sector_size_bits = ffs(vol->sector_size) - 1; in parse_ntfs_boot_sector() 760 ntfs_debug("vol->sector_size = %i (0x%x)", vol->sector_size, in parse_ntfs_boot_sector() 761 vol->sector_size); in parse_ntfs_boot_sector() 764 if (vol->sector_size < vol->sb->s_blocksize) { in parse_ntfs_boot_sector() 767 "supported. Sorry.", vol->sector_size, in parse_ntfs_boot_sector() 777 vol->cluster_size = vol->sector_size << sectors_per_cluster_bits; in parse_ntfs_boot_sector() 784 if (vol->cluster_size < vol->sector_size) { in parse_ntfs_boot_sector() 787 "Sorry.", vol->cluster_size, vol->sector_size); in parse_ntfs_boot_sector() 823 if (vol->mft_record_size < vol->sector_size) { in parse_ntfs_boot_sector() [all …]
|
D | volume.h | 44 u16 sector_size; /* in bytes */ member
|
/fs/xfs/ |
D | xfs_mount.c | 271 unsigned int sector_size; in xfs_readsb() local 288 sector_size = xfs_getsize_buftarg(mp->m_ddev_targp); in xfs_readsb() 299 BTOBB(sector_size), XBF_NO_IOACCT, &bp, in xfs_readsb() 329 if (sector_size > sbp->sb_sectsize) { in xfs_readsb() 332 sector_size, sbp->sb_sectsize); in xfs_readsb() 343 sector_size = sbp->sb_sectsize; in xfs_readsb()
|
/fs/ocfs2/ |
D | super.c | 116 int sector_size, 726 int *sector_size, in ocfs2_sb_probe() argument 737 *sector_size = bdev_logical_block_size(sb->s_bdev); in ocfs2_sb_probe() 738 if (*sector_size > OCFS2_MAX_BLOCKSIZE) { in ocfs2_sb_probe() 740 *sector_size, OCFS2_MAX_BLOCKSIZE); in ocfs2_sb_probe() 746 if (*sector_size < OCFS2_MIN_BLOCKSIZE) in ocfs2_sb_probe() 747 *sector_size = OCFS2_MIN_BLOCKSIZE; in ocfs2_sb_probe() 750 status = ocfs2_get_sector(sb, bh, 0, *sector_size); in ocfs2_sb_probe() 781 for (blksize = *sector_size; in ocfs2_sb_probe() 975 int status, sector_size; in ocfs2_fill_super() local [all …]
|
/fs/ |
D | dax.c | 1050 unsigned short sector_size = bdev_logical_block_size(bdev); in dax_range_is_aligned() local 1052 if (!IS_ALIGNED(offset, sector_size)) in dax_range_is_aligned() 1054 if (!IS_ALIGNED(length, sector_size)) in dax_range_is_aligned()
|
/fs/btrfs/ |
D | dev-replace.c | 161 dev_replace->tgtdev->sector_size = fs_info->sectorsize; in btrfs_init_dev_replace() 245 device->sector_size = fs_info->sectorsize; in btrfs_init_dev_replace_tgtdev()
|
D | volumes.h | 111 u32 sector_size; member
|
D | ctree.h | 1422 BTRFS_SETGET_FUNCS(device_sector_size, struct btrfs_dev_item, sector_size, 32); 1439 sector_size, 32); 1465 BTRFS_SETGET_FUNCS(chunk_sector_size, struct btrfs_chunk, sector_size, 32); 1486 sector_size, 32);
|
/fs/fat/ |
D | inode.c | 1436 if (get_unaligned_le16(&b->sector_size)) in fat_bpb_is_zero() 1466 bpb->fat_sector_size = get_unaligned_le16(&b->sector_size); in fat_read_bpb()
|