Searched refs:sds (Results 1 – 3 of 3) sorted by relevance
668 struct sched_domain_shared *sds = NULL; in update_top_cache_domain() local677 sds = sd->shared; in update_top_cache_domain()683 rcu_assign_pointer(per_cpu(sd_llc_shared, cpu), sds); in update_top_cache_domain()1489 if (atomic_read(&(*per_cpu_ptr(sdd->sds, cpu))->ref)) in claim_allocations()1490 *per_cpu_ptr(sdd->sds, cpu) = NULL; in claim_allocations()1634 sd->shared = *per_cpu_ptr(sdd->sds, sd_id); in sd_init()2077 sdd->sds = alloc_percpu(struct sched_domain_shared *); in __sdt_alloc()2078 if (!sdd->sds) in __sdt_alloc()2091 struct sched_domain_shared *sds; in __sdt_alloc() local2102 sds = kzalloc_node(sizeof(struct sched_domain_shared), in __sdt_alloc()[all …]
6661 struct sched_domain_shared *sds; in set_idle_cores() local6663 sds = rcu_dereference(per_cpu(sd_llc_shared, cpu)); in set_idle_cores()6664 if (sds) in set_idle_cores()6665 WRITE_ONCE(sds->has_idle_cores, val); in set_idle_cores()6670 struct sched_domain_shared *sds; in test_idle_cores() local6672 sds = rcu_dereference(per_cpu(sd_llc_shared, cpu)); in test_idle_cores()6673 if (sds) in test_idle_cores()6674 return READ_ONCE(sds->has_idle_cores); in test_idle_cores()8978 static inline void init_sd_lb_stats(struct sd_lb_stats *sds) in init_sd_lb_stats() argument8987 *sds = (struct sd_lb_stats){ in init_sd_lb_stats()[all …]
655 struct perf_sample_data sds[3]; member663 struct bpf_trace_sample_data *sds; in BPF_CALL_5() local674 sds = this_cpu_ptr(&bpf_trace_sds); in BPF_CALL_5()677 if (WARN_ON_ONCE(nest_level > ARRAY_SIZE(sds->sds))) { in BPF_CALL_5()682 sd = &sds->sds[nest_level - 1]; in BPF_CALL_5()743 if (WARN_ON_ONCE(nest_level > ARRAY_SIZE(bpf_misc_sds.sds))) { in bpf_event_output()747 sd = this_cpu_ptr(&bpf_misc_sds.sds[nest_level - 1]); in bpf_event_output()