Lines Matching refs:bdi
48 struct backing_dev_info *bdi = m->private; in bdi_debug_stats_show() local
49 struct bdi_writeback *wb = &bdi->wb; in bdi_debug_stats_show()
100 !list_empty(&bdi->bdi_list), bdi->wb.state); in bdi_debug_stats_show()
107 static void bdi_debug_register(struct backing_dev_info *bdi, const char *name) in bdi_debug_register() argument
109 bdi->debug_dir = debugfs_create_dir(name, bdi_debug_root); in bdi_debug_register()
111 debugfs_create_file("stats", 0444, bdi->debug_dir, bdi, in bdi_debug_register()
115 static void bdi_debug_unregister(struct backing_dev_info *bdi) in bdi_debug_unregister() argument
117 debugfs_remove_recursive(bdi->debug_dir); in bdi_debug_unregister()
123 static inline void bdi_debug_register(struct backing_dev_info *bdi, in bdi_debug_register() argument
127 static inline void bdi_debug_unregister(struct backing_dev_info *bdi) in bdi_debug_unregister() argument
136 struct backing_dev_info *bdi = dev_get_drvdata(dev); in read_ahead_kb_store() local
144 bdi->ra_pages = read_ahead_kb >> (PAGE_SHIFT - 10); in read_ahead_kb_store()
155 struct backing_dev_info *bdi = dev_get_drvdata(dev); \
161 BDI_SHOW(read_ahead_kb, K(bdi->ra_pages))
166 struct backing_dev_info *bdi = dev_get_drvdata(dev); in min_ratio_store() local
174 ret = bdi_set_min_ratio(bdi, ratio); in min_ratio_store()
180 BDI_SHOW(min_ratio, bdi->min_ratio)
185 struct backing_dev_info *bdi = dev_get_drvdata(dev); in max_ratio_store() local
193 ret = bdi_set_max_ratio(bdi, ratio); in max_ratio_store()
199 BDI_SHOW(max_ratio, bdi->max_ratio)
233 static int bdi_init(struct backing_dev_info *bdi);
280 static int wb_init(struct bdi_writeback *wb, struct backing_dev_info *bdi, in wb_init() argument
287 if (wb != &bdi->wb) in wb_init()
288 bdi_get(bdi); in wb_init()
289 wb->bdi = bdi; in wb_init()
325 if (wb != &bdi->wb) in wb_init()
326 bdi_put(bdi); in wb_init()
366 if (wb != &wb->bdi->wb) in wb_exit()
367 bdi_put(wb->bdi); in wb_exit()
396 mutex_lock(&wb->bdi->cgwb_release_mutex); in cgwb_release_workfn()
401 mutex_unlock(&wb->bdi->cgwb_release_mutex); in cgwb_release_workfn()
422 WARN_ON(!radix_tree_delete(&wb->bdi->cgwb_tree, wb->memcg_css->id)); in cgwb_kill()
435 static int cgwb_create(struct backing_dev_info *bdi, in cgwb_create() argument
454 wb = radix_tree_lookup(&bdi->cgwb_tree, memcg_css->id); in cgwb_create()
470 ret = wb_init(wb, bdi, gfp); in cgwb_create()
495 if (test_bit(WB_registered, &bdi->wb.state) && in cgwb_create()
498 ret = radix_tree_insert(&bdi->cgwb_tree, memcg_css->id, wb); in cgwb_create()
500 list_add_tail_rcu(&wb->bdi_node, &bdi->wb_list); in cgwb_create()
552 struct bdi_writeback *wb_get_lookup(struct backing_dev_info *bdi, in wb_get_lookup() argument
558 return &bdi->wb; in wb_get_lookup()
561 wb = radix_tree_lookup(&bdi->cgwb_tree, memcg_css->id); in wb_get_lookup()
585 struct bdi_writeback *wb_get_create(struct backing_dev_info *bdi, in wb_get_create() argument
594 return &bdi->wb; in wb_get_create()
597 wb = wb_get_lookup(bdi, memcg_css); in wb_get_create()
598 } while (!wb && !cgwb_create(bdi, memcg_css, gfp)); in wb_get_create()
603 static int cgwb_bdi_init(struct backing_dev_info *bdi) in cgwb_bdi_init() argument
607 INIT_RADIX_TREE(&bdi->cgwb_tree, GFP_ATOMIC); in cgwb_bdi_init()
608 mutex_init(&bdi->cgwb_release_mutex); in cgwb_bdi_init()
609 init_rwsem(&bdi->wb_switch_rwsem); in cgwb_bdi_init()
611 ret = wb_init(&bdi->wb, bdi, GFP_KERNEL); in cgwb_bdi_init()
613 bdi->wb.memcg_css = &root_mem_cgroup->css; in cgwb_bdi_init()
614 bdi->wb.blkcg_css = blkcg_root_css; in cgwb_bdi_init()
619 static void cgwb_bdi_unregister(struct backing_dev_info *bdi) in cgwb_bdi_unregister() argument
625 WARN_ON(test_bit(WB_registered, &bdi->wb.state)); in cgwb_bdi_unregister()
628 radix_tree_for_each_slot(slot, &bdi->cgwb_tree, &iter, 0) in cgwb_bdi_unregister()
632 mutex_lock(&bdi->cgwb_release_mutex); in cgwb_bdi_unregister()
634 while (!list_empty(&bdi->wb_list)) { in cgwb_bdi_unregister()
635 wb = list_first_entry(&bdi->wb_list, struct bdi_writeback, in cgwb_bdi_unregister()
642 mutex_unlock(&bdi->cgwb_release_mutex); in cgwb_bdi_unregister()
680 static void cgwb_bdi_register(struct backing_dev_info *bdi) in cgwb_bdi_register() argument
683 list_add_tail_rcu(&bdi->wb.bdi_node, &bdi->wb_list); in cgwb_bdi_register()
704 static int cgwb_bdi_init(struct backing_dev_info *bdi) in cgwb_bdi_init() argument
706 return wb_init(&bdi->wb, bdi, GFP_KERNEL); in cgwb_bdi_init()
709 static void cgwb_bdi_unregister(struct backing_dev_info *bdi) { } in cgwb_bdi_unregister() argument
711 static void cgwb_bdi_register(struct backing_dev_info *bdi) in cgwb_bdi_register() argument
713 list_add_tail_rcu(&bdi->wb.bdi_node, &bdi->wb_list); in cgwb_bdi_register()
723 static int bdi_init(struct backing_dev_info *bdi) in bdi_init() argument
727 bdi->dev = NULL; in bdi_init()
729 kref_init(&bdi->refcnt); in bdi_init()
730 bdi->min_ratio = 0; in bdi_init()
731 bdi->max_ratio = 100; in bdi_init()
732 bdi->max_prop_frac = FPROP_FRAC_BASE; in bdi_init()
733 INIT_LIST_HEAD(&bdi->bdi_list); in bdi_init()
734 INIT_LIST_HEAD(&bdi->wb_list); in bdi_init()
735 init_waitqueue_head(&bdi->wb_waitq); in bdi_init()
737 ret = cgwb_bdi_init(bdi); in bdi_init()
744 struct backing_dev_info *bdi; in bdi_alloc() local
746 bdi = kzalloc_node(sizeof(*bdi), GFP_KERNEL, node_id); in bdi_alloc()
747 if (!bdi) in bdi_alloc()
750 if (bdi_init(bdi)) { in bdi_alloc()
751 kfree(bdi); in bdi_alloc()
754 bdi->capabilities = BDI_CAP_WRITEBACK | BDI_CAP_WRITEBACK_ACCT; in bdi_alloc()
755 bdi->ra_pages = VM_READAHEAD_PAGES; in bdi_alloc()
756 bdi->io_pages = VM_READAHEAD_PAGES; in bdi_alloc()
757 return bdi; in bdi_alloc()
765 struct backing_dev_info *bdi; in bdi_lookup_rb_node() local
771 bdi = rb_entry(parent, struct backing_dev_info, rb_node); in bdi_lookup_rb_node()
773 if (bdi->id > id) in bdi_lookup_rb_node()
775 else if (bdi->id < id) in bdi_lookup_rb_node()
795 struct backing_dev_info *bdi = NULL; in bdi_get_by_id() local
801 bdi = rb_entry(*p, struct backing_dev_info, rb_node); in bdi_get_by_id()
802 bdi_get(bdi); in bdi_get_by_id()
806 return bdi; in bdi_get_by_id()
809 int bdi_register_va(struct backing_dev_info *bdi, const char *fmt, va_list args) in bdi_register_va() argument
814 if (bdi->dev) /* The driver needs to use separate queues per device */ in bdi_register_va()
817 vsnprintf(bdi->dev_name, sizeof(bdi->dev_name), fmt, args); in bdi_register_va()
818 dev = device_create(bdi_class, NULL, MKDEV(0, 0), bdi, bdi->dev_name); in bdi_register_va()
822 cgwb_bdi_register(bdi); in bdi_register_va()
823 bdi->dev = dev; in bdi_register_va()
825 bdi_debug_register(bdi, dev_name(dev)); in bdi_register_va()
826 set_bit(WB_registered, &bdi->wb.state); in bdi_register_va()
830 bdi->id = ++bdi_id_cursor; in bdi_register_va()
832 p = bdi_lookup_rb_node(bdi->id, &parent); in bdi_register_va()
833 rb_link_node(&bdi->rb_node, parent, p); in bdi_register_va()
834 rb_insert_color(&bdi->rb_node, &bdi_tree); in bdi_register_va()
836 list_add_tail_rcu(&bdi->bdi_list, &bdi_list); in bdi_register_va()
840 trace_writeback_bdi_register(bdi); in bdi_register_va()
844 int bdi_register(struct backing_dev_info *bdi, const char *fmt, ...) in bdi_register() argument
850 ret = bdi_register_va(bdi, fmt, args); in bdi_register()
856 void bdi_set_owner(struct backing_dev_info *bdi, struct device *owner) in bdi_set_owner() argument
858 WARN_ON_ONCE(bdi->owner); in bdi_set_owner()
859 bdi->owner = owner; in bdi_set_owner()
866 static void bdi_remove_from_list(struct backing_dev_info *bdi) in bdi_remove_from_list() argument
869 rb_erase(&bdi->rb_node, &bdi_tree); in bdi_remove_from_list()
870 list_del_rcu(&bdi->bdi_list); in bdi_remove_from_list()
876 void bdi_unregister(struct backing_dev_info *bdi) in bdi_unregister() argument
879 bdi_remove_from_list(bdi); in bdi_unregister()
880 wb_shutdown(&bdi->wb); in bdi_unregister()
881 cgwb_bdi_unregister(bdi); in bdi_unregister()
887 if (bdi->min_ratio) in bdi_unregister()
888 bdi_set_min_ratio(bdi, 0); in bdi_unregister()
890 if (bdi->dev) { in bdi_unregister()
891 bdi_debug_unregister(bdi); in bdi_unregister()
892 device_unregister(bdi->dev); in bdi_unregister()
893 bdi->dev = NULL; in bdi_unregister()
896 if (bdi->owner) { in bdi_unregister()
897 put_device(bdi->owner); in bdi_unregister()
898 bdi->owner = NULL; in bdi_unregister()
904 struct backing_dev_info *bdi = in release_bdi() local
907 if (test_bit(WB_registered, &bdi->wb.state)) in release_bdi()
908 bdi_unregister(bdi); in release_bdi()
909 WARN_ON_ONCE(bdi->dev); in release_bdi()
910 wb_exit(&bdi->wb); in release_bdi()
911 kfree(bdi); in release_bdi()
914 void bdi_put(struct backing_dev_info *bdi) in bdi_put() argument
916 kref_put(&bdi->refcnt, release_bdi); in bdi_put()
920 const char *bdi_dev_name(struct backing_dev_info *bdi) in bdi_dev_name() argument
922 if (!bdi || !bdi->dev) in bdi_dev_name()
924 return bdi->dev_name; in bdi_dev_name()
934 void clear_bdi_congested(struct backing_dev_info *bdi, int sync) in clear_bdi_congested() argument
940 if (test_and_clear_bit(bit, &bdi->wb.congested)) in clear_bdi_congested()
948 void set_bdi_congested(struct backing_dev_info *bdi, int sync) in set_bdi_congested() argument
953 if (!test_and_set_bit(bit, &bdi->wb.congested)) in set_bdi_congested()