• Home
  • Raw
  • Download

Lines Matching refs:c

145 int bch_journal_read(struct cache_set *c, struct list_head *list)  in bch_journal_read()  argument
159 for_each_cache(ca, c, iter) { in bch_journal_read()
259 c->journal.seq = list_entry(list->prev, in bch_journal_read()
267 void bch_journal_mark(struct cache_set *c, struct list_head *list) in bch_journal_mark() argument
272 struct journal *j = &c->journal; in bch_journal_mark()
301 if (!__bch_extent_invalid(c, k)) { in bch_journal_mark()
305 if (ptr_available(c, k, j)) in bch_journal_mark()
306 atomic_inc(&PTR_BUCKET(c, k, j)->pin); in bch_journal_mark()
308 bch_initial_mark_key(c, 0, k); in bch_journal_mark()
387 static void btree_flush_write(struct cache_set *c) in btree_flush_write() argument
398 for_each_cached_btree(b, c, i) in btree_flush_write()
402 else if (journal_pin_cmp(c, in btree_flush_write()
486 static void journal_reclaim(struct cache_set *c) in journal_reclaim() argument
488 struct bkey *k = &c->journal.key; in journal_reclaim()
494 while (!atomic_read(&fifo_front(&c->journal.pin))) in journal_reclaim()
495 fifo_pop(&c->journal.pin, p); in journal_reclaim()
497 last_seq = last_seq(&c->journal); in journal_reclaim()
501 for_each_cache(ca, c, iter) { in journal_reclaim()
510 for_each_cache(ca, c, iter) in journal_reclaim()
513 if (c->journal.blocks_free) in journal_reclaim()
521 for_each_cache(ca, c, iter) { in journal_reclaim()
531 bucket_to_sector(c, ca->sb.d[ja->cur_idx]), in journal_reclaim()
538 c->journal.blocks_free = c->sb.bucket_size >> c->block_bits; in journal_reclaim()
541 if (!journal_full(&c->journal)) in journal_reclaim()
542 __closure_wake_up(&c->journal.wait); in journal_reclaim()
573 cache_set_err_on(bio->bi_status, w->c, "journal io error"); in journal_write_endio()
574 closure_put(&w->c->journal.io); in journal_write_endio()
592 struct cache_set *c = container_of(cl, struct cache_set, journal.io); in journal_write_unlock() local
594 c->journal.io_in_flight = 0; in journal_write_unlock()
595 spin_unlock(&c->journal.lock); in journal_write_unlock()
599 __releases(c->journal.lock) in journal_write_unlocked()
601 struct cache_set *c = container_of(cl, struct cache_set, journal.io); in journal_write_unlocked() local
603 struct journal_write *w = c->journal.cur; in journal_write_unlocked()
604 struct bkey *k = &c->journal.key; in journal_write_unlocked()
605 unsigned i, sectors = set_blocks(w->data, block_bytes(c)) * in journal_write_unlocked()
606 c->sb.block_size; in journal_write_unlocked()
615 } else if (journal_full(&c->journal)) { in journal_write_unlocked()
616 journal_reclaim(c); in journal_write_unlocked()
617 spin_unlock(&c->journal.lock); in journal_write_unlocked()
619 btree_flush_write(c); in journal_write_unlocked()
624 c->journal.blocks_free -= set_blocks(w->data, block_bytes(c)); in journal_write_unlocked()
626 w->data->btree_level = c->root->level; in journal_write_unlocked()
628 bkey_copy(&w->data->btree_root, &c->root->key); in journal_write_unlocked()
629 bkey_copy(&w->data->uuid_bucket, &c->uuid_bucket); in journal_write_unlocked()
631 for_each_cache(ca, c, i) in journal_write_unlocked()
634 w->data->magic = jset_magic(&c->sb); in journal_write_unlocked()
636 w->data->last_seq = last_seq(&c->journal); in journal_write_unlocked()
640 ca = PTR_CACHE(c, k, i); in journal_write_unlocked()
667 atomic_dec_bug(&fifo_back(&c->journal.pin)); in journal_write_unlocked()
668 bch_journal_next(&c->journal); in journal_write_unlocked()
669 journal_reclaim(c); in journal_write_unlocked()
671 spin_unlock(&c->journal.lock); in journal_write_unlocked()
681 struct cache_set *c = container_of(cl, struct cache_set, journal.io); in journal_write() local
683 spin_lock(&c->journal.lock); in journal_write()
687 static void journal_try_write(struct cache_set *c) in journal_try_write() argument
688 __releases(c->journal.lock) in journal_try_write()
690 struct closure *cl = &c->journal.io; in journal_try_write()
691 struct journal_write *w = c->journal.cur; in journal_try_write()
695 if (!c->journal.io_in_flight) { in journal_try_write()
696 c->journal.io_in_flight = 1; in journal_try_write()
697 closure_call(cl, journal_write_unlocked, NULL, &c->cl); in journal_try_write()
699 spin_unlock(&c->journal.lock); in journal_try_write()
703 static struct journal_write *journal_wait_for_write(struct cache_set *c, in journal_wait_for_write() argument
712 spin_lock(&c->journal.lock); in journal_wait_for_write()
715 struct journal_write *w = c->journal.cur; in journal_wait_for_write()
718 block_bytes(c)) * c->sb.block_size; in journal_wait_for_write()
721 c->journal.blocks_free * c->sb.block_size, in journal_wait_for_write()
726 closure_wait(&c->journal.wait, &cl); in journal_wait_for_write()
728 if (!journal_full(&c->journal)) { in journal_wait_for_write()
730 trace_bcache_journal_entry_full(c); in journal_wait_for_write()
740 journal_try_write(c); /* unlocks */ in journal_wait_for_write()
743 trace_bcache_journal_full(c); in journal_wait_for_write()
745 journal_reclaim(c); in journal_wait_for_write()
746 spin_unlock(&c->journal.lock); in journal_wait_for_write()
748 btree_flush_write(c); in journal_wait_for_write()
752 spin_lock(&c->journal.lock); in journal_wait_for_write()
759 struct cache_set *c = container_of(to_delayed_work(work), in journal_write_work() local
762 spin_lock(&c->journal.lock); in journal_write_work()
763 if (c->journal.cur->dirty) in journal_write_work()
764 journal_try_write(c); in journal_write_work()
766 spin_unlock(&c->journal.lock); in journal_write_work()
775 atomic_t *bch_journal(struct cache_set *c, in bch_journal() argument
782 if (!CACHE_SYNC(&c->sb)) in bch_journal()
785 w = journal_wait_for_write(c, bch_keylist_nkeys(keys)); in bch_journal()
790 ret = &fifo_back(&c->journal.pin); in bch_journal()
795 journal_try_write(c); in bch_journal()
798 schedule_delayed_work(&c->journal.work, in bch_journal()
799 msecs_to_jiffies(c->journal_delay_ms)); in bch_journal()
800 spin_unlock(&c->journal.lock); in bch_journal()
802 spin_unlock(&c->journal.lock); in bch_journal()
809 void bch_journal_meta(struct cache_set *c, struct closure *cl) in bch_journal_meta() argument
816 ref = bch_journal(c, &keys, cl); in bch_journal_meta()
821 void bch_journal_free(struct cache_set *c) in bch_journal_free() argument
823 free_pages((unsigned long) c->journal.w[1].data, JSET_BITS); in bch_journal_free()
824 free_pages((unsigned long) c->journal.w[0].data, JSET_BITS); in bch_journal_free()
825 free_fifo(&c->journal.pin); in bch_journal_free()
828 int bch_journal_alloc(struct cache_set *c) in bch_journal_alloc() argument
830 struct journal *j = &c->journal; in bch_journal_alloc()
835 c->journal_delay_ms = 100; in bch_journal_alloc()
837 j->w[0].c = c; in bch_journal_alloc()
838 j->w[1].c = c; in bch_journal_alloc()