• Home
  • Raw
  • Download

Lines Matching full:cs

213 static inline struct cpuset *parent_cs(struct cpuset *cs)  in parent_cs()  argument
215 return css_cs(cs->css.parent); in parent_cs()
220 struct cpuset *cs = task_cs(p); in inc_dl_tasks_cs() local
222 cs->nr_deadline_tasks++; in inc_dl_tasks_cs()
227 struct cpuset *cs = task_cs(p); in dec_dl_tasks_cs() local
229 cs->nr_deadline_tasks--; in dec_dl_tasks_cs()
245 static inline bool is_cpuset_online(struct cpuset *cs) in is_cpuset_online() argument
247 return test_bit(CS_ONLINE, &cs->flags) && !css_is_dying(&cs->css); in is_cpuset_online()
250 static inline int is_cpu_exclusive(const struct cpuset *cs) in is_cpu_exclusive() argument
252 return test_bit(CS_CPU_EXCLUSIVE, &cs->flags); in is_cpu_exclusive()
255 static inline int is_mem_exclusive(const struct cpuset *cs) in is_mem_exclusive() argument
257 return test_bit(CS_MEM_EXCLUSIVE, &cs->flags); in is_mem_exclusive()
260 static inline int is_mem_hardwall(const struct cpuset *cs) in is_mem_hardwall() argument
262 return test_bit(CS_MEM_HARDWALL, &cs->flags); in is_mem_hardwall()
265 static inline int is_sched_load_balance(const struct cpuset *cs) in is_sched_load_balance() argument
267 return test_bit(CS_SCHED_LOAD_BALANCE, &cs->flags); in is_sched_load_balance()
270 static inline int is_memory_migrate(const struct cpuset *cs) in is_memory_migrate() argument
272 return test_bit(CS_MEMORY_MIGRATE, &cs->flags); in is_memory_migrate()
275 static inline int is_spread_page(const struct cpuset *cs) in is_spread_page() argument
277 return test_bit(CS_SPREAD_PAGE, &cs->flags); in is_spread_page()
280 static inline int is_spread_slab(const struct cpuset *cs) in is_spread_slab() argument
282 return test_bit(CS_SPREAD_SLAB, &cs->flags); in is_spread_slab()
285 static inline int is_partition_root(const struct cpuset *cs) in is_partition_root() argument
287 return cs->partition_root_state > 0; in is_partition_root()
408 static void guarantee_online_cpus(struct cpuset *cs, struct cpumask *pmask) in guarantee_online_cpus() argument
410 while (!cpumask_intersects(cs->effective_cpus, cpu_online_mask)) { in guarantee_online_cpus()
411 cs = parent_cs(cs); in guarantee_online_cpus()
412 if (unlikely(!cs)) { in guarantee_online_cpus()
424 cpumask_and(pmask, cs->effective_cpus, cpu_online_mask); in guarantee_online_cpus()
438 static void guarantee_online_mems(struct cpuset *cs, nodemask_t *pmask) in guarantee_online_mems() argument
440 while (!nodes_intersects(cs->effective_mems, node_states[N_MEMORY])) in guarantee_online_mems()
441 cs = parent_cs(cs); in guarantee_online_mems()
442 nodes_and(*pmask, cs->effective_mems, node_states[N_MEMORY]); in guarantee_online_mems()
450 static void cpuset_update_task_spread_flag(struct cpuset *cs, in cpuset_update_task_spread_flag() argument
453 if (is_spread_page(cs)) in cpuset_update_task_spread_flag()
458 if (is_spread_slab(cs)) in cpuset_update_task_spread_flag()
482 * @cs: the cpuset that have cpumasks to be allocated.
488 static inline int alloc_cpumasks(struct cpuset *cs, struct tmpmasks *tmp) in alloc_cpumasks() argument
492 if (cs) { in alloc_cpumasks()
493 pmask1 = &cs->cpus_allowed; in alloc_cpumasks()
494 pmask2 = &cs->effective_cpus; in alloc_cpumasks()
495 pmask3 = &cs->subparts_cpus; in alloc_cpumasks()
496 pmask4 = &cs->cpus_requested; in alloc_cpumasks()
512 if (cs && !zalloc_cpumask_var(pmask4, GFP_KERNEL)) in alloc_cpumasks()
528 * @cs: the cpuset that have cpumasks to be free.
531 static inline void free_cpumasks(struct cpuset *cs, struct tmpmasks *tmp) in free_cpumasks() argument
533 if (cs) { in free_cpumasks()
534 free_cpumask_var(cs->cpus_allowed); in free_cpumasks()
535 free_cpumask_var(cs->cpus_requested); in free_cpumasks()
536 free_cpumask_var(cs->effective_cpus); in free_cpumasks()
537 free_cpumask_var(cs->subparts_cpus); in free_cpumasks()
548 * @cs: the cpuset that the trial cpuset duplicates
550 static struct cpuset *alloc_trial_cpuset(struct cpuset *cs) in alloc_trial_cpuset() argument
554 trial = kmemdup(cs, sizeof(*cs), GFP_KERNEL); in alloc_trial_cpuset()
563 cpumask_copy(trial->cpus_allowed, cs->cpus_allowed); in alloc_trial_cpuset()
564 cpumask_copy(trial->cpus_requested, cs->cpus_requested); in alloc_trial_cpuset()
565 cpumask_copy(trial->effective_cpus, cs->effective_cpus); in alloc_trial_cpuset()
571 * @cs: the cpuset to be freed
573 static inline void free_cpuset(struct cpuset *cs) in free_cpuset() argument
575 free_cpumasks(cs, NULL); in free_cpuset()
576 kfree(cs); in free_cpuset()
947 static void dl_update_tasks_root_domain(struct cpuset *cs) in dl_update_tasks_root_domain() argument
952 if (cs->nr_deadline_tasks == 0) in dl_update_tasks_root_domain()
955 css_task_iter_start(&cs->css, 0, &it); in dl_update_tasks_root_domain()
965 struct cpuset *cs = NULL; in dl_rebuild_rd_accounting() local
980 cpuset_for_each_descendant_pre(cs, pos_css, &top_cpuset) { in dl_rebuild_rd_accounting()
982 if (cpumask_empty(cs->effective_cpus)) { in dl_rebuild_rd_accounting()
987 css_get(&cs->css); in dl_rebuild_rd_accounting()
991 dl_update_tasks_root_domain(cs); in dl_rebuild_rd_accounting()
994 css_put(&cs->css); in dl_rebuild_rd_accounting()
1025 struct cpuset *cs; in rebuild_sched_domains_locked() local
1051 cpuset_for_each_descendant_pre(cs, pos_css, &top_cpuset) { in rebuild_sched_domains_locked()
1052 if (!is_partition_root(cs)) { in rebuild_sched_domains_locked()
1056 if (!cpumask_subset(cs->effective_cpus, in rebuild_sched_domains_locked()
1088 * @cs: the cpuset in which each task's cpus_allowed mask needs to be changed
1090 * Iterate through each task of @cs updating its cpus_allowed to the
1094 static void update_tasks_cpumask(struct cpuset *cs) in update_tasks_cpumask() argument
1098 bool top_cs = cs == &top_cpuset; in update_tasks_cpumask()
1100 css_task_iter_start(&cs->css, 0, &it); in update_tasks_cpumask()
1108 set_cpus_allowed_ptr(task, cs->effective_cpus); in update_tasks_cpumask()
1116 * @cs: the cpuset the need to recompute the new effective_cpus mask
1125 struct cpuset *cs, struct cpuset *parent) in compute_effective_cpumask() argument
1130 cpumask_and(new_cpus, new_cpus, cs->cpus_requested); in compute_effective_cpumask()
1133 cpumask_and(new_cpus, cs->cpus_requested, in compute_effective_cpumask()
1359 * @cs: the cpuset to consider
1369 static void update_cpumasks_hier(struct cpuset *cs, struct tmpmasks *tmp) in update_cpumasks_hier() argument
1377 cpuset_for_each_descendant_pre(cp, pos_css, cs) { in update_cpumasks_hier()
1410 * for cs already in update_cpumask(). We should also call in update_cpumasks_hier()
1415 if ((cp != cs) && new_prs) { in update_cpumasks_hier()
1520 * @cs: Current cpuset
1523 static void update_sibling_cpumasks(struct cpuset *parent, struct cpuset *cs, in update_sibling_cpumasks() argument
1541 if (sibling == cs) in update_sibling_cpumasks()
1558 * @cs: the cpuset to consider
1562 static int update_cpumask(struct cpuset *cs, struct cpuset *trialcs, in update_cpumask() argument
1569 if (cs == &top_cpuset) in update_cpumask()
1593 if (cpumask_equal(cs->cpus_requested, trialcs->cpus_requested)) in update_cpumask()
1596 retval = validate_change(cs, trialcs); in update_cpumask()
1610 if (cs->partition_root_state) { in update_cpumask()
1614 if (update_parent_subparts_cpumask(cs, partcmd_update, in update_cpumask()
1620 cpumask_copy(cs->cpus_allowed, trialcs->cpus_allowed); in update_cpumask()
1621 cpumask_copy(cs->cpus_requested, trialcs->cpus_requested); in update_cpumask()
1626 if (cs->nr_subparts_cpus) { in update_cpumask()
1627 cpumask_and(cs->subparts_cpus, cs->subparts_cpus, cs->cpus_allowed); in update_cpumask()
1628 cs->nr_subparts_cpus = cpumask_weight(cs->subparts_cpus); in update_cpumask()
1632 update_cpumasks_hier(cs, &tmp); in update_cpumask()
1634 if (cs->partition_root_state) { in update_cpumask()
1635 struct cpuset *parent = parent_cs(cs); in update_cpumask()
1642 update_sibling_cpumasks(parent, cs, &tmp); in update_cpumask()
1727 * @cs: the cpuset in which each task's mems_allowed mask needs to be changed
1729 * Iterate through each task of @cs updating its mems_allowed to the
1733 static void update_tasks_nodemask(struct cpuset *cs) in update_tasks_nodemask() argument
1739 cpuset_being_rebound = cs; /* causes mpol_dup() rebind */ in update_tasks_nodemask()
1741 guarantee_online_mems(cs, &newmems); in update_tasks_nodemask()
1753 css_task_iter_start(&cs->css, 0, &it); in update_tasks_nodemask()
1764 migrate = is_memory_migrate(cs); in update_tasks_nodemask()
1766 mpol_rebind_mm(mm, &cs->mems_allowed); in update_tasks_nodemask()
1768 cpuset_migrate_mm(mm, &cs->old_mems_allowed, &newmems); in update_tasks_nodemask()
1776 * cs->old_mems_allowed. in update_tasks_nodemask()
1778 cs->old_mems_allowed = newmems; in update_tasks_nodemask()
1786 * @cs: the cpuset to consider
1796 static void update_nodemasks_hier(struct cpuset *cs, nodemask_t *new_mems) in update_nodemasks_hier() argument
1802 cpuset_for_each_descendant_pre(cp, pos_css, cs) { in update_nodemasks_hier()
1848 * Will take tasklist_lock, scan tasklist for tasks in cpuset cs,
1852 static int update_nodemask(struct cpuset *cs, struct cpuset *trialcs, in update_nodemask() argument
1861 if (cs == &top_cpuset) { in update_nodemask()
1886 if (nodes_equal(cs->mems_allowed, trialcs->mems_allowed)) { in update_nodemask()
1890 retval = validate_change(cs, trialcs); in update_nodemask()
1895 cs->mems_allowed = trialcs->mems_allowed; in update_nodemask()
1899 update_nodemasks_hier(cs, &trialcs->mems_allowed); in update_nodemask()
1915 static int update_relax_domain_level(struct cpuset *cs, s64 val) in update_relax_domain_level() argument
1922 if (val != cs->relax_domain_level) { in update_relax_domain_level()
1923 cs->relax_domain_level = val; in update_relax_domain_level()
1924 if (!cpumask_empty(cs->cpus_allowed) && in update_relax_domain_level()
1925 is_sched_load_balance(cs)) in update_relax_domain_level()
1934 * @cs: the cpuset in which each task's spread flags needs to be changed
1936 * Iterate through each task of @cs updating its spread flags. As this
1940 static void update_tasks_flags(struct cpuset *cs) in update_tasks_flags() argument
1945 css_task_iter_start(&cs->css, 0, &it); in update_tasks_flags()
1947 cpuset_update_task_spread_flag(cs, task); in update_tasks_flags()
1954 * cs: the cpuset to update
1960 static int update_flag(cpuset_flagbits_t bit, struct cpuset *cs, in update_flag() argument
1968 trialcs = alloc_trial_cpuset(cs); in update_flag()
1977 err = validate_change(cs, trialcs); in update_flag()
1981 balance_flag_changed = (is_sched_load_balance(cs) != in update_flag()
1984 spread_flag_changed = ((is_spread_slab(cs) != is_spread_slab(trialcs)) in update_flag()
1985 || (is_spread_page(cs) != is_spread_page(trialcs))); in update_flag()
1988 cs->flags = trialcs->flags; in update_flag()
1995 update_tasks_flags(cs); in update_flag()
2003 * cs: the cpuset to update
2008 static int update_prstate(struct cpuset *cs, int new_prs) in update_prstate() argument
2010 int err, old_prs = cs->partition_root_state; in update_prstate()
2011 struct cpuset *parent = parent_cs(cs); in update_prstate()
2034 if (cpumask_empty(cs->cpus_allowed)) in update_prstate()
2037 err = update_flag(CS_CPU_EXCLUSIVE, cs, 1); in update_prstate()
2041 err = update_parent_subparts_cpumask(cs, partcmd_enable, in update_prstate()
2044 update_flag(CS_CPU_EXCLUSIVE, cs, 0); in update_prstate()
2053 update_flag(CS_CPU_EXCLUSIVE, cs, 0); in update_prstate()
2058 err = update_parent_subparts_cpumask(cs, partcmd_disable, in update_prstate()
2064 update_flag(CS_CPU_EXCLUSIVE, cs, 0); in update_prstate()
2070 update_sibling_cpumasks(parent, cs, &tmpmask); in update_prstate()
2076 cs->partition_root_state = new_prs; in update_prstate()
2187 static void reset_migrate_dl_data(struct cpuset *cs) in reset_migrate_dl_data() argument
2189 cs->nr_migrate_dl_tasks = 0; in reset_migrate_dl_data()
2190 cs->sum_migrate_dl_bw = 0; in reset_migrate_dl_data()
2197 struct cpuset *cs, *oldcs; in cpuset_can_attach() local
2204 cs = css_cs(css); in cpuset_can_attach()
2211 (cpumask_empty(cs->cpus_allowed) || nodes_empty(cs->mems_allowed))) in cpuset_can_attach()
2223 cs->nr_migrate_dl_tasks++; in cpuset_can_attach()
2224 cs->sum_migrate_dl_bw += task->dl.dl_bw; in cpuset_can_attach()
2228 if (!cs->nr_migrate_dl_tasks) in cpuset_can_attach()
2231 if (!cpumask_intersects(oldcs->effective_cpus, cs->effective_cpus)) { in cpuset_can_attach()
2232 int cpu = cpumask_any_and(cpu_active_mask, cs->effective_cpus); in cpuset_can_attach()
2235 reset_migrate_dl_data(cs); in cpuset_can_attach()
2240 ret = dl_bw_alloc(cpu, cs->sum_migrate_dl_bw); in cpuset_can_attach()
2242 reset_migrate_dl_data(cs); in cpuset_can_attach()
2252 cs->attach_in_progress++; in cpuset_can_attach()
2262 struct cpuset *cs; in cpuset_cancel_attach() local
2265 cs = css_cs(css); in cpuset_cancel_attach()
2268 cs->attach_in_progress--; in cpuset_cancel_attach()
2269 if (!cs->attach_in_progress) in cpuset_cancel_attach()
2272 if (cs->nr_migrate_dl_tasks) { in cpuset_cancel_attach()
2273 int cpu = cpumask_any(cs->effective_cpus); in cpuset_cancel_attach()
2275 dl_bw_free(cpu, cs->sum_migrate_dl_bw); in cpuset_cancel_attach()
2276 reset_migrate_dl_data(cs); in cpuset_cancel_attach()
2296 struct cpuset *cs; in cpuset_attach() local
2300 cs = css_cs(css); in cpuset_attach()
2306 if (cs == &top_cpuset) in cpuset_attach()
2309 guarantee_online_cpus(cs, cpus_attach); in cpuset_attach()
2311 guarantee_online_mems(cs, &cpuset_attach_nodemask_to); in cpuset_attach()
2321 cpuset_update_task_spread_flag(cs, task); in cpuset_attach()
2328 cpuset_attach_nodemask_to = cs->effective_mems; in cpuset_attach()
2343 if (is_memory_migrate(cs)) in cpuset_attach()
2351 cs->old_mems_allowed = cpuset_attach_nodemask_to; in cpuset_attach()
2353 if (cs->nr_migrate_dl_tasks) { in cpuset_attach()
2354 cs->nr_deadline_tasks += cs->nr_migrate_dl_tasks; in cpuset_attach()
2355 oldcs->nr_deadline_tasks -= cs->nr_migrate_dl_tasks; in cpuset_attach()
2356 reset_migrate_dl_data(cs); in cpuset_attach()
2359 cs->attach_in_progress--; in cpuset_attach()
2360 if (!cs->attach_in_progress) in cpuset_attach()
2390 struct cpuset *cs = css_cs(css); in cpuset_write_u64() local
2396 if (!is_cpuset_online(cs)) { in cpuset_write_u64()
2403 retval = update_flag(CS_CPU_EXCLUSIVE, cs, val); in cpuset_write_u64()
2406 retval = update_flag(CS_MEM_EXCLUSIVE, cs, val); in cpuset_write_u64()
2409 retval = update_flag(CS_MEM_HARDWALL, cs, val); in cpuset_write_u64()
2412 retval = update_flag(CS_SCHED_LOAD_BALANCE, cs, val); in cpuset_write_u64()
2415 retval = update_flag(CS_MEMORY_MIGRATE, cs, val); in cpuset_write_u64()
2421 retval = update_flag(CS_SPREAD_PAGE, cs, val); in cpuset_write_u64()
2424 retval = update_flag(CS_SPREAD_SLAB, cs, val); in cpuset_write_u64()
2439 struct cpuset *cs = css_cs(css); in cpuset_write_s64() local
2445 if (!is_cpuset_online(cs)) in cpuset_write_s64()
2450 retval = update_relax_domain_level(cs, val); in cpuset_write_s64()
2468 struct cpuset *cs = css_cs(of_css(of)); in cpuset_write_resmask() local
2475 * CPU or memory hotunplug may leave @cs w/o any execution in cpuset_write_resmask()
2480 * As writes to "cpus" or "mems" may restore @cs's execution in cpuset_write_resmask()
2489 * protection is okay as we check whether @cs is online after in cpuset_write_resmask()
2493 css_get(&cs->css); in cpuset_write_resmask()
2499 if (!is_cpuset_online(cs)) in cpuset_write_resmask()
2502 trialcs = alloc_trial_cpuset(cs); in cpuset_write_resmask()
2510 retval = update_cpumask(cs, trialcs, buf); in cpuset_write_resmask()
2513 retval = update_nodemask(cs, trialcs, buf); in cpuset_write_resmask()
2525 css_put(&cs->css); in cpuset_write_resmask()
2540 struct cpuset *cs = css_cs(seq_css(sf)); in cpuset_common_seq_show() local
2548 seq_printf(sf, "%*pbl\n", cpumask_pr_args(cs->cpus_requested)); in cpuset_common_seq_show()
2551 seq_printf(sf, "%*pbl\n", nodemask_pr_args(&cs->mems_allowed)); in cpuset_common_seq_show()
2554 seq_printf(sf, "%*pbl\n", cpumask_pr_args(cs->effective_cpus)); in cpuset_common_seq_show()
2557 seq_printf(sf, "%*pbl\n", nodemask_pr_args(&cs->effective_mems)); in cpuset_common_seq_show()
2560 seq_printf(sf, "%*pbl\n", cpumask_pr_args(cs->subparts_cpus)); in cpuset_common_seq_show()
2572 struct cpuset *cs = css_cs(css); in cpuset_read_u64() local
2576 return is_cpu_exclusive(cs); in cpuset_read_u64()
2578 return is_mem_exclusive(cs); in cpuset_read_u64()
2580 return is_mem_hardwall(cs); in cpuset_read_u64()
2582 return is_sched_load_balance(cs); in cpuset_read_u64()
2584 return is_memory_migrate(cs); in cpuset_read_u64()
2588 return fmeter_getrate(&cs->fmeter); in cpuset_read_u64()
2590 return is_spread_page(cs); in cpuset_read_u64()
2592 return is_spread_slab(cs); in cpuset_read_u64()
2603 struct cpuset *cs = css_cs(css); in cpuset_read_s64() local
2607 return cs->relax_domain_level; in cpuset_read_s64()
2618 struct cpuset *cs = css_cs(seq_css(seq)); in sched_partition_show() local
2620 switch (cs->partition_root_state) { in sched_partition_show()
2637 struct cpuset *cs = css_cs(of_css(of)); in sched_partition_write() local
2653 css_get(&cs->css); in sched_partition_write()
2656 if (!is_cpuset_online(cs)) in sched_partition_write()
2659 retval = update_prstate(cs, val); in sched_partition_write()
2663 css_put(&cs->css); in sched_partition_write()
2835 struct cpuset *cs; in cpuset_css_alloc() local
2840 cs = kzalloc(sizeof(*cs), GFP_KERNEL); in cpuset_css_alloc()
2841 if (!cs) in cpuset_css_alloc()
2844 if (alloc_cpumasks(cs, NULL)) { in cpuset_css_alloc()
2845 kfree(cs); in cpuset_css_alloc()
2849 set_bit(CS_SCHED_LOAD_BALANCE, &cs->flags); in cpuset_css_alloc()
2850 nodes_clear(cs->mems_allowed); in cpuset_css_alloc()
2851 nodes_clear(cs->effective_mems); in cpuset_css_alloc()
2852 fmeter_init(&cs->fmeter); in cpuset_css_alloc()
2853 cs->relax_domain_level = -1; in cpuset_css_alloc()
2855 return &cs->css; in cpuset_css_alloc()
2860 struct cpuset *cs = css_cs(css); in cpuset_css_online() local
2861 struct cpuset *parent = parent_cs(cs); in cpuset_css_online()
2871 set_bit(CS_ONLINE, &cs->flags); in cpuset_css_online()
2873 set_bit(CS_SPREAD_PAGE, &cs->flags); in cpuset_css_online()
2875 set_bit(CS_SPREAD_SLAB, &cs->flags); in cpuset_css_online()
2881 cpumask_copy(cs->effective_cpus, parent->effective_cpus); in cpuset_css_online()
2882 cs->effective_mems = parent->effective_mems; in cpuset_css_online()
2883 cs->use_parent_ecpus = true; in cpuset_css_online()
2914 cs->mems_allowed = parent->mems_allowed; in cpuset_css_online()
2915 cs->effective_mems = parent->mems_allowed; in cpuset_css_online()
2916 cpumask_copy(cs->cpus_allowed, parent->cpus_allowed); in cpuset_css_online()
2917 cpumask_copy(cs->cpus_requested, parent->cpus_requested); in cpuset_css_online()
2918 cpumask_copy(cs->effective_cpus, parent->cpus_allowed); in cpuset_css_online()
2939 struct cpuset *cs = css_cs(css); in cpuset_css_offline() local
2944 if (is_partition_root(cs)) in cpuset_css_offline()
2945 update_prstate(cs, 0); in cpuset_css_offline()
2948 is_sched_load_balance(cs)) in cpuset_css_offline()
2949 update_flag(CS_SCHED_LOAD_BALANCE, cs, 0); in cpuset_css_offline()
2951 if (cs->use_parent_ecpus) { in cpuset_css_offline()
2952 struct cpuset *parent = parent_cs(cs); in cpuset_css_offline()
2954 cs->use_parent_ecpus = false; in cpuset_css_offline()
2959 clear_bit(CS_ONLINE, &cs->flags); in cpuset_css_offline()
2967 struct cpuset *cs = css_cs(css); in cpuset_css_free() local
2969 free_cpuset(cs); in cpuset_css_free()
3056 static void remove_tasks_in_empty_cpuset(struct cpuset *cs) in remove_tasks_in_empty_cpuset() argument
3064 parent = parent_cs(cs); in remove_tasks_in_empty_cpuset()
3069 if (cgroup_transfer_tasks(parent->css.cgroup, cs->css.cgroup)) { in remove_tasks_in_empty_cpuset()
3071 pr_cont_cgroup_name(cs->css.cgroup); in remove_tasks_in_empty_cpuset()
3077 hotplug_update_tasks_legacy(struct cpuset *cs, in hotplug_update_tasks_legacy() argument
3084 cpumask_copy(cs->cpus_allowed, new_cpus); in hotplug_update_tasks_legacy()
3085 cpumask_copy(cs->effective_cpus, new_cpus); in hotplug_update_tasks_legacy()
3086 cs->mems_allowed = *new_mems; in hotplug_update_tasks_legacy()
3087 cs->effective_mems = *new_mems; in hotplug_update_tasks_legacy()
3094 if (cpus_updated && !cpumask_empty(cs->cpus_allowed)) in hotplug_update_tasks_legacy()
3095 update_tasks_cpumask(cs); in hotplug_update_tasks_legacy()
3096 if (mems_updated && !nodes_empty(cs->mems_allowed)) in hotplug_update_tasks_legacy()
3097 update_tasks_nodemask(cs); in hotplug_update_tasks_legacy()
3099 is_empty = cpumask_empty(cs->cpus_allowed) || in hotplug_update_tasks_legacy()
3100 nodes_empty(cs->mems_allowed); in hotplug_update_tasks_legacy()
3110 remove_tasks_in_empty_cpuset(cs); in hotplug_update_tasks_legacy()
3116 hotplug_update_tasks(struct cpuset *cs, in hotplug_update_tasks() argument
3121 cpumask_copy(new_cpus, parent_cs(cs)->effective_cpus); in hotplug_update_tasks()
3123 *new_mems = parent_cs(cs)->effective_mems; in hotplug_update_tasks()
3126 cpumask_copy(cs->effective_cpus, new_cpus); in hotplug_update_tasks()
3127 cs->effective_mems = *new_mems; in hotplug_update_tasks()
3131 update_tasks_cpumask(cs); in hotplug_update_tasks()
3133 update_tasks_nodemask(cs); in hotplug_update_tasks()
3145 * @cs: cpuset in interest
3148 * Compare @cs's cpu and mem masks against top_cpuset and if some have gone
3149 * offline, update @cs accordingly. If @cs ends up with no CPU or memory,
3152 static void cpuset_hotplug_update_tasks(struct cpuset *cs, struct tmpmasks *tmp) in cpuset_hotplug_update_tasks() argument
3160 wait_event(cpuset_attach_wq, cs->attach_in_progress == 0); in cpuset_hotplug_update_tasks()
3168 if (cs->attach_in_progress) { in cpuset_hotplug_update_tasks()
3173 parent = parent_cs(cs); in cpuset_hotplug_update_tasks()
3174 compute_effective_cpumask(&new_cpus, cs, parent); in cpuset_hotplug_update_tasks()
3175 nodes_and(new_mems, cs->mems_allowed, parent->effective_mems); in cpuset_hotplug_update_tasks()
3177 if (cs->nr_subparts_cpus) in cpuset_hotplug_update_tasks()
3182 cpumask_andnot(&new_cpus, &new_cpus, cs->subparts_cpus); in cpuset_hotplug_update_tasks()
3184 if (!tmp || !cs->partition_root_state) in cpuset_hotplug_update_tasks()
3192 if (is_partition_root(cs) && (cpumask_empty(&new_cpus) || in cpuset_hotplug_update_tasks()
3194 if (cs->nr_subparts_cpus) { in cpuset_hotplug_update_tasks()
3196 cs->nr_subparts_cpus = 0; in cpuset_hotplug_update_tasks()
3197 cpumask_clear(cs->subparts_cpus); in cpuset_hotplug_update_tasks()
3199 compute_effective_cpumask(&new_cpus, cs, parent); in cpuset_hotplug_update_tasks()
3210 update_parent_subparts_cpumask(cs, partcmd_disable, in cpuset_hotplug_update_tasks()
3213 cs->partition_root_state = PRS_ERROR; in cpuset_hotplug_update_tasks()
3225 ((cs->partition_root_state == PRS_ERROR) || in cpuset_hotplug_update_tasks()
3227 update_parent_subparts_cpumask(cs, partcmd_update, NULL, tmp)) in cpuset_hotplug_update_tasks()
3231 cpus_updated = !cpumask_equal(&new_cpus, cs->effective_cpus); in cpuset_hotplug_update_tasks()
3232 mems_updated = !nodes_equal(new_mems, cs->effective_mems); in cpuset_hotplug_update_tasks()
3235 hotplug_update_tasks(cs, &new_cpus, &new_mems, in cpuset_hotplug_update_tasks()
3238 hotplug_update_tasks_legacy(cs, &new_cpus, &new_mems, in cpuset_hotplug_update_tasks()
3332 struct cpuset *cs; in cpuset_hotplug_workfn() local
3336 cpuset_for_each_descendant_pre(cs, pos_css, &top_cpuset) { in cpuset_hotplug_workfn()
3337 if (cs == &top_cpuset || !css_tryget_online(&cs->css)) in cpuset_hotplug_workfn()
3341 cpuset_hotplug_update_tasks(cs, ptmp); in cpuset_hotplug_workfn()
3344 css_put(&cs->css); in cpuset_hotplug_workfn()
3457 * But we used cs && cs->cpus_allowed lockless and thus can in cpuset_cpus_allowed_fallback()
3519 static struct cpuset *nearest_hardwall_ancestor(struct cpuset *cs) in nearest_hardwall_ancestor() argument
3521 while (!(is_mem_exclusive(cs) || is_mem_hardwall(cs)) && parent_cs(cs)) in nearest_hardwall_ancestor()
3522 cs = parent_cs(cs); in nearest_hardwall_ancestor()
3523 return cs; in nearest_hardwall_ancestor()
3568 struct cpuset *cs; /* current cpuset ancestors */ in __cpuset_node_allowed() local
3592 cs = nearest_hardwall_ancestor(task_cs(current)); in __cpuset_node_allowed()
3593 allowed = node_isset(node, cs->mems_allowed); in __cpuset_node_allowed()