• Home
  • Raw
  • Download

Lines Matching refs:part

92 static int build_block_map(struct partition *part, int block_no)  in build_block_map()  argument
94 struct block *block = &part->blocks[block_no]; in build_block_map()
97 block->offset = part->block_size * block_no; in build_block_map()
99 if (le16_to_cpu(part->header_cache[0]) != RFD_MAGIC) { in build_block_map()
106 for (i=0; i<part->data_sectors_per_block; i++) { in build_block_map()
109 entry = le16_to_cpu(part->header_cache[HEADER_MAP_OFFSET + i]); in build_block_map()
122 if (entry >= part->sector_count) { in build_block_map()
126 part->mbd.mtd->name, block_no, i, entry); in build_block_map()
130 if (part->sector_map[entry] != -1) { in build_block_map()
133 part->mbd.mtd->name, entry); in build_block_map()
134 part->errors = 1; in build_block_map()
138 part->sector_map[entry] = block->offset + in build_block_map()
139 (i + part->header_sectors_per_block) * SECTOR_SIZE; in build_block_map()
144 if (block->free_sectors == part->data_sectors_per_block) in build_block_map()
145 part->reserved_block = block_no; in build_block_map()
150 static int scan_header(struct partition *part) in scan_header() argument
157 sectors_per_block = part->block_size / SECTOR_SIZE; in scan_header()
158 part->total_blocks = (u32)part->mbd.mtd->size / part->block_size; in scan_header()
160 if (part->total_blocks < 2) in scan_header()
164 part->header_sectors_per_block = in scan_header()
168 part->data_sectors_per_block = sectors_per_block - in scan_header()
169 part->header_sectors_per_block; in scan_header()
171 part->header_size = (HEADER_MAP_OFFSET + in scan_header()
172 part->data_sectors_per_block) * sizeof(u16); in scan_header()
174 part->cylinders = (part->data_sectors_per_block * in scan_header()
175 (part->total_blocks - 1) - 1) / SECTORS_PER_TRACK; in scan_header()
177 part->sector_count = part->cylinders * SECTORS_PER_TRACK; in scan_header()
179 part->current_block = -1; in scan_header()
180 part->reserved_block = -1; in scan_header()
181 part->is_reclaiming = 0; in scan_header()
183 part->header_cache = kmalloc(part->header_size, GFP_KERNEL); in scan_header()
184 if (!part->header_cache) in scan_header()
187 part->blocks = kcalloc(part->total_blocks, sizeof(struct block), in scan_header()
189 if (!part->blocks) in scan_header()
192 part->sector_map = vmalloc(part->sector_count * sizeof(u_long)); in scan_header()
193 if (!part->sector_map) { in scan_header()
195 "sector map", part->mbd.mtd->name); in scan_header()
199 for (i=0; i<part->sector_count; i++) in scan_header()
200 part->sector_map[i] = -1; in scan_header()
202 for (i=0, blocks_found=0; i<part->total_blocks; i++) { in scan_header()
203 rc = mtd_read(part->mbd.mtd, i * part->block_size, in scan_header()
204 part->header_size, &retlen, in scan_header()
205 (u_char *)part->header_cache); in scan_header()
207 if (!rc && retlen != part->header_size) in scan_header()
213 if (!build_block_map(part, i)) in scan_header()
219 part->mbd.mtd->name); in scan_header()
224 if (part->reserved_block == -1) { in scan_header()
226 part->mbd.mtd->name); in scan_header()
228 part->errors = 1; in scan_header()
234 vfree(part->sector_map); in scan_header()
235 kfree(part->header_cache); in scan_header()
236 kfree(part->blocks); in scan_header()
243 struct partition *part = (struct partition*)dev; in rfd_ftl_readsect() local
248 if (sector >= part->sector_count) in rfd_ftl_readsect()
251 addr = part->sector_map[sector]; in rfd_ftl_readsect()
253 rc = mtd_read(part->mbd.mtd, addr, SECTOR_SIZE, &retlen, in rfd_ftl_readsect()
260 "0x%lx\n", part->mbd.mtd->name, addr); in rfd_ftl_readsect()
271 struct partition *part; in erase_callback() local
276 part = (struct partition*)erase->priv; in erase_callback()
278 i = (u32)erase->addr / part->block_size; in erase_callback()
279 if (i >= part->total_blocks || part->blocks[i].offset != erase->addr || in erase_callback()
282 "on '%s'\n", (unsigned long long)erase->addr, part->mbd.mtd->name); in erase_callback()
289 part->mbd.mtd->name, erase->state); in erase_callback()
291 part->blocks[i].state = BLOCK_FAILED; in erase_callback()
292 part->blocks[i].free_sectors = 0; in erase_callback()
293 part->blocks[i].used_sectors = 0; in erase_callback()
302 part->blocks[i].state = BLOCK_ERASED; in erase_callback()
303 part->blocks[i].free_sectors = part->data_sectors_per_block; in erase_callback()
304 part->blocks[i].used_sectors = 0; in erase_callback()
305 part->blocks[i].erases++; in erase_callback()
307 rc = mtd_write(part->mbd.mtd, part->blocks[i].offset, sizeof(magic), in erase_callback()
316 part->mbd.mtd->name, in erase_callback()
317 part->blocks[i].offset); in erase_callback()
318 part->blocks[i].state = BLOCK_FAILED; in erase_callback()
321 part->blocks[i].state = BLOCK_OK; in erase_callback()
326 static int erase_block(struct partition *part, int block) in erase_block() argument
335 erase->mtd = part->mbd.mtd; in erase_block()
337 erase->addr = part->blocks[block].offset; in erase_block()
338 erase->len = part->block_size; in erase_block()
339 erase->priv = (u_long)part; in erase_block()
341 part->blocks[block].state = BLOCK_ERASING; in erase_block()
342 part->blocks[block].free_sectors = 0; in erase_block()
344 rc = mtd_erase(part->mbd.mtd, erase); in erase_block()
349 (unsigned long long)erase->len, part->mbd.mtd->name); in erase_block()
357 static int move_block_contents(struct partition *part, int block_no, u_long *old_sector) in move_block_contents() argument
364 part->is_reclaiming = 1; in move_block_contents()
370 map = kmalloc(part->header_size, GFP_KERNEL); in move_block_contents()
374 rc = mtd_read(part->mbd.mtd, part->blocks[block_no].offset, in move_block_contents()
375 part->header_size, &retlen, (u_char *)map); in move_block_contents()
377 if (!rc && retlen != part->header_size) in move_block_contents()
382 "0x%lx\n", part->mbd.mtd->name, in move_block_contents()
383 part->blocks[block_no].offset); in move_block_contents()
388 for (i=0; i<part->data_sectors_per_block; i++) { in move_block_contents()
400 if (entry >= part->sector_count) in move_block_contents()
403 addr = part->blocks[block_no].offset + in move_block_contents()
404 (i + part->header_sectors_per_block) * SECTOR_SIZE; in move_block_contents()
408 if (!part->blocks[block_no].used_sectors--) { in move_block_contents()
409 rc = erase_block(part, block_no); in move_block_contents()
414 rc = mtd_read(part->mbd.mtd, addr, SECTOR_SIZE, &retlen, in move_block_contents()
423 part->mbd.mtd->name); in move_block_contents()
428 rc = rfd_ftl_writesect((struct mtd_blktrans_dev*)part, in move_block_contents()
440 part->is_reclaiming = 0; in move_block_contents()
445 static int reclaim_block(struct partition *part, u_long *old_sector) in reclaim_block() argument
451 mtd_sync(part->mbd.mtd); in reclaim_block()
456 old_sector_block = *old_sector / part->block_size; in reclaim_block()
460 for (block=0; block<part->total_blocks; block++) { in reclaim_block()
463 if (block == part->reserved_block) in reclaim_block()
471 if (part->blocks[block].free_sectors) in reclaim_block()
474 this_score = part->blocks[block].used_sectors; in reclaim_block()
480 if (part->blocks[block].used_sectors == in reclaim_block()
481 part->data_sectors_per_block) in reclaim_block()
485 this_score += part->blocks[block].erases; in reclaim_block()
496 part->current_block = -1; in reclaim_block()
497 part->reserved_block = best_block; in reclaim_block()
501 part->blocks[best_block].used_sectors, in reclaim_block()
502 part->blocks[best_block].free_sectors); in reclaim_block()
504 if (part->blocks[best_block].used_sectors) in reclaim_block()
505 rc = move_block_contents(part, best_block, old_sector); in reclaim_block()
507 rc = erase_block(part, best_block); in reclaim_block()
517 static int find_free_block(struct partition *part) in find_free_block() argument
521 block = part->current_block == -1 ? in find_free_block()
522 jiffies % part->total_blocks : part->current_block; in find_free_block()
526 if (part->blocks[block].free_sectors && in find_free_block()
527 block != part->reserved_block) in find_free_block()
530 if (part->blocks[block].state == BLOCK_UNUSED) in find_free_block()
531 erase_block(part, block); in find_free_block()
533 if (++block >= part->total_blocks) in find_free_block()
541 static int find_writable_block(struct partition *part, u_long *old_sector) in find_writable_block() argument
546 block = find_free_block(part); in find_writable_block()
549 if (!part->is_reclaiming) { in find_writable_block()
550 rc = reclaim_block(part, old_sector); in find_writable_block()
554 block = find_free_block(part); in find_writable_block()
563 rc = mtd_read(part->mbd.mtd, part->blocks[block].offset, in find_writable_block()
564 part->header_size, &retlen, in find_writable_block()
565 (u_char *)part->header_cache); in find_writable_block()
567 if (!rc && retlen != part->header_size) in find_writable_block()
572 "0x%lx\n", part->mbd.mtd->name, in find_writable_block()
573 part->blocks[block].offset); in find_writable_block()
577 part->current_block = block; in find_writable_block()
583 static int mark_sector_deleted(struct partition *part, u_long old_addr) in mark_sector_deleted() argument
590 block = old_addr / part->block_size; in mark_sector_deleted()
591 offset = (old_addr % part->block_size) / SECTOR_SIZE - in mark_sector_deleted()
592 part->header_sectors_per_block; in mark_sector_deleted()
594 addr = part->blocks[block].offset + in mark_sector_deleted()
596 rc = mtd_write(part->mbd.mtd, addr, sizeof(del), &retlen, in mark_sector_deleted()
604 "0x%lx\n", part->mbd.mtd->name, addr); in mark_sector_deleted()
608 if (block == part->current_block) in mark_sector_deleted()
609 part->header_cache[offset + HEADER_MAP_OFFSET] = del; in mark_sector_deleted()
611 part->blocks[block].used_sectors--; in mark_sector_deleted()
613 if (!part->blocks[block].used_sectors && in mark_sector_deleted()
614 !part->blocks[block].free_sectors) in mark_sector_deleted()
615 rc = erase_block(part, block); in mark_sector_deleted()
621 static int find_free_sector(const struct partition *part, const struct block *block) in find_free_sector() argument
625 i = stop = part->data_sectors_per_block - block->free_sectors; in find_free_sector()
628 if (le16_to_cpu(part->header_cache[HEADER_MAP_OFFSET + i]) in find_free_sector()
632 if (++i == part->data_sectors_per_block) in find_free_sector()
642 struct partition *part = (struct partition*)dev; in do_writesect() local
650 if (part->current_block == -1 || in do_writesect()
651 !part->blocks[part->current_block].free_sectors) { in do_writesect()
653 rc = find_writable_block(part, old_addr); in do_writesect()
658 block = &part->blocks[part->current_block]; in do_writesect()
660 i = find_free_sector(part, block); in do_writesect()
667 addr = (i + part->header_sectors_per_block) * SECTOR_SIZE + in do_writesect()
669 rc = mtd_write(part->mbd.mtd, addr, SECTOR_SIZE, &retlen, in do_writesect()
677 part->mbd.mtd->name, addr); in do_writesect()
682 part->sector_map[sector] = addr; in do_writesect()
686 part->header_cache[i + HEADER_MAP_OFFSET] = entry; in do_writesect()
689 rc = mtd_write(part->mbd.mtd, addr, sizeof(entry), &retlen, in do_writesect()
697 part->mbd.mtd->name, addr); in do_writesect()
710 struct partition *part = (struct partition*)dev; in rfd_ftl_writesect() local
717 if (part->reserved_block == -1) { in rfd_ftl_writesect()
722 if (sector >= part->sector_count) { in rfd_ftl_writesect()
727 old_addr = part->sector_map[sector]; in rfd_ftl_writesect()
740 part->sector_map[sector] = -1; in rfd_ftl_writesect()
743 rc = mark_sector_deleted(part, old_addr); in rfd_ftl_writesect()
751 struct partition *part = (struct partition*)dev; in rfd_ftl_getgeo() local
755 geo->cylinders = part->cylinders; in rfd_ftl_getgeo()
762 struct partition *part; in rfd_ftl_add_mtd() local
767 part = kzalloc(sizeof(struct partition), GFP_KERNEL); in rfd_ftl_add_mtd()
768 if (!part) in rfd_ftl_add_mtd()
771 part->mbd.mtd = mtd; in rfd_ftl_add_mtd()
774 part->block_size = block_size; in rfd_ftl_add_mtd()
780 part->block_size = mtd->erasesize; in rfd_ftl_add_mtd()
783 if (scan_header(part) == 0) { in rfd_ftl_add_mtd()
784 part->mbd.size = part->sector_count; in rfd_ftl_add_mtd()
785 part->mbd.tr = tr; in rfd_ftl_add_mtd()
786 part->mbd.devnum = -1; in rfd_ftl_add_mtd()
788 part->mbd.readonly = 1; in rfd_ftl_add_mtd()
789 else if (part->errors) { in rfd_ftl_add_mtd()
792 part->mbd.readonly = 1; in rfd_ftl_add_mtd()
798 if (!add_mtd_blktrans_dev((void*)part)) in rfd_ftl_add_mtd()
802 kfree(part); in rfd_ftl_add_mtd()
807 struct partition *part = (struct partition*)dev; in rfd_ftl_remove_dev() local
810 for (i=0; i<part->total_blocks; i++) { in rfd_ftl_remove_dev()
812 part->mbd.mtd->name, i, part->blocks[i].erases); in rfd_ftl_remove_dev()
816 vfree(part->sector_map); in rfd_ftl_remove_dev()
817 kfree(part->header_cache); in rfd_ftl_remove_dev()
818 kfree(part->blocks); in rfd_ftl_remove_dev()