Home
last modified time | relevance | path

Searched refs:cpumask (Results 1 – 25 of 216) sorted by relevance

123456789

/arch/x86/include/asm/uv/
Duv.h6 struct cpumask;
16 extern const struct cpumask *uv_flush_tlb_others(const struct cpumask *cpumask,
27 static inline const struct cpumask *
28 uv_flush_tlb_others(const struct cpumask *cpumask, struct mm_struct *mm, in uv_flush_tlb_others() argument
30 { return cpumask; } in uv_flush_tlb_others()
/arch/x86/kernel/apic/
Dapic_flat_64.c39 static const struct cpumask *flat_target_cpus(void) in flat_target_cpus()
44 static void flat_vector_allocation_domain(int cpu, struct cpumask *retmask) in flat_vector_allocation_domain()
87 static void flat_send_IPI_mask(const struct cpumask *cpumask, int vector) in flat_send_IPI_mask() argument
89 unsigned long mask = cpumask_bits(cpumask)[0]; in flat_send_IPI_mask()
95 flat_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) in flat_send_IPI_mask_allbutself() argument
97 unsigned long mask = cpumask_bits(cpumask)[0]; in flat_send_IPI_mask_allbutself()
264 static const struct cpumask *physflat_target_cpus(void) in physflat_target_cpus()
269 static void physflat_vector_allocation_domain(int cpu, struct cpumask *retmask) in physflat_vector_allocation_domain()
275 static void physflat_send_IPI_mask(const struct cpumask *cpumask, int vector) in physflat_send_IPI_mask() argument
277 default_send_IPI_mask_sequence_phys(cpumask, vector); in physflat_send_IPI_mask()
[all …]
Dx2apic_phys.c39 __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) in __x2apic_send_IPI_mask()
59 static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask()
65 x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask_allbutself()
80 static unsigned int x2apic_cpu_mask_to_apicid(const struct cpumask *cpumask) in x2apic_cpu_mask_to_apicid() argument
86 int cpu = cpumask_first(cpumask); in x2apic_cpu_mask_to_apicid()
95 x2apic_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in x2apic_cpu_mask_to_apicid_and() argument
96 const struct cpumask *andmask) in x2apic_cpu_mask_to_apicid_and()
104 for_each_cpu_and(cpu, cpumask, andmask) { in x2apic_cpu_mask_to_apicid_and()
Dipi.c22 void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector) in default_send_IPI_mask_sequence_phys()
40 void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, in default_send_IPI_mask_allbutself_phys()
61 void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, in default_send_IPI_mask_sequence_logical()
81 void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, in default_send_IPI_mask_allbutself_logical()
104 void default_send_IPI_mask_logical(const struct cpumask *cpumask, int vector) in default_send_IPI_mask_logical() argument
106 unsigned long mask = cpumask_bits(cpumask)[0]; in default_send_IPI_mask_logical()
Dbigsmp_32.c29 static const struct cpumask *bigsmp_target_cpus(void) in bigsmp_target_cpus()
109 static unsigned int bigsmp_cpu_mask_to_apicid(const struct cpumask *cpumask) in bigsmp_cpu_mask_to_apicid() argument
111 int cpu = cpumask_first(cpumask); in bigsmp_cpu_mask_to_apicid()
118 static unsigned int bigsmp_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in bigsmp_cpu_mask_to_apicid_and() argument
119 const struct cpumask *andmask) in bigsmp_cpu_mask_to_apicid_and()
127 for_each_cpu_and(cpu, cpumask, andmask) { in bigsmp_cpu_mask_to_apicid_and()
139 static inline void bigsmp_send_IPI_mask(const struct cpumask *mask, int vector) in bigsmp_send_IPI_mask()
180 static void bigsmp_vector_allocation_domain(int cpu, struct cpumask *retmask) in bigsmp_vector_allocation_domain()
Dx2apic_cluster.c28 __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) in __x2apic_send_IPI_mask()
30 struct cpumask *cpus_in_cluster_ptr; in __x2apic_send_IPI_mask()
31 struct cpumask *ipi_mask_ptr; in __x2apic_send_IPI_mask()
78 static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask()
84 x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask_allbutself()
99 static unsigned int x2apic_cpu_mask_to_apicid(const struct cpumask *cpumask) in x2apic_cpu_mask_to_apicid() argument
105 int cpu = cpumask_first(cpumask); in x2apic_cpu_mask_to_apicid()
114 x2apic_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in x2apic_cpu_mask_to_apicid_and() argument
115 const struct cpumask *andmask) in x2apic_cpu_mask_to_apicid_and()
123 for_each_cpu_and(cpu, cpumask, andmask) { in x2apic_cpu_mask_to_apicid_and()
Dapic_numachip.c75 static const struct cpumask *numachip_target_cpus(void) in numachip_target_cpus()
80 static void numachip_vector_allocation_domain(int cpu, struct cpumask *retmask) in numachip_vector_allocation_domain()
119 static void numachip_send_IPI_mask(const struct cpumask *mask, int vector) in numachip_send_IPI_mask()
127 static void numachip_send_IPI_mask_allbutself(const struct cpumask *mask, in numachip_send_IPI_mask_allbutself()
160 static unsigned int numachip_cpu_mask_to_apicid(const struct cpumask *cpumask) in numachip_cpu_mask_to_apicid() argument
168 cpu = cpumask_first(cpumask); in numachip_cpu_mask_to_apicid()
176 numachip_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in numachip_cpu_mask_to_apicid_and() argument
177 const struct cpumask *andmask) in numachip_cpu_mask_to_apicid_and()
185 for_each_cpu_and(cpu, cpumask, andmask) { in numachip_cpu_mask_to_apicid_and()
Dsummit_32.c55 static inline void summit_send_IPI_mask(const struct cpumask *mask, int vector) in summit_send_IPI_mask()
177 static const struct cpumask *summit_target_cpus(void) in summit_target_cpus()
266 static unsigned int summit_cpu_mask_to_apicid(const struct cpumask *cpumask) in summit_cpu_mask_to_apicid() argument
274 for_each_cpu(cpu, cpumask) { in summit_cpu_mask_to_apicid()
287 static unsigned int summit_cpu_mask_to_apicid_and(const struct cpumask *inmask, in summit_cpu_mask_to_apicid_and()
288 const struct cpumask *andmask) in summit_cpu_mask_to_apicid_and()
291 cpumask_var_t cpumask; in summit_cpu_mask_to_apicid_and() local
293 if (!alloc_cpumask_var(&cpumask, GFP_ATOMIC)) in summit_cpu_mask_to_apicid_and()
296 cpumask_and(cpumask, inmask, andmask); in summit_cpu_mask_to_apicid_and()
297 cpumask_and(cpumask, cpumask, cpu_online_mask); in summit_cpu_mask_to_apicid_and()
[all …]
Dapic_noop.c34 static void noop_send_IPI_mask(const struct cpumask *cpumask, int vector) { } in noop_send_IPI_mask() argument
35 static void noop_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) { } in noop_send_IPI_mask_allbutself() argument
87 static const struct cpumask *noop_target_cpus(void) in noop_target_cpus()
103 static void noop_vector_allocation_domain(int cpu, struct cpumask *retmask) in noop_vector_allocation_domain()
Des7000_32.c397 static void es7000_vector_allocation_domain(int cpu, struct cpumask *retmask) in es7000_vector_allocation_domain()
423 static void es7000_send_IPI_mask(const struct cpumask *mask, int vector) in es7000_send_IPI_mask()
443 static const struct cpumask *target_cpus_cluster(void) in target_cpus_cluster()
448 static const struct cpumask *es7000_target_cpus(void) in es7000_target_cpus()
543 static unsigned int es7000_cpu_mask_to_apicid(const struct cpumask *cpumask) in es7000_cpu_mask_to_apicid() argument
551 for_each_cpu(cpu, cpumask) { in es7000_cpu_mask_to_apicid()
566 es7000_cpu_mask_to_apicid_and(const struct cpumask *inmask, in es7000_cpu_mask_to_apicid_and()
567 const struct cpumask *andmask) in es7000_cpu_mask_to_apicid_and()
570 cpumask_var_t cpumask; in es7000_cpu_mask_to_apicid_and() local
572 if (!alloc_cpumask_var(&cpumask, GFP_ATOMIC)) in es7000_cpu_mask_to_apicid_and()
[all …]
/arch/m32r/kernel/
Dsmp.c80 static void send_IPI_mask(const struct cpumask *, int, int);
151 cpumask_t cpumask; in smp_flush_cache_all() local
155 cpumask_copy(&cpumask, cpu_online_mask); in smp_flush_cache_all()
156 cpumask_clear_cpu(smp_processor_id(), &cpumask); in smp_flush_cache_all()
158 mask=cpumask_bits(&cpumask); in smp_flush_cache_all()
160 send_IPI_mask(&cpumask, INVALIDATE_CACHE_IPI, 0); in smp_flush_cache_all()
370 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, in flush_tlb_others() argument
388 BUG_ON(cpumask_empty(&cpumask)); in flush_tlb_others()
390 BUG_ON(cpumask_test_cpu(smp_processor_id(), &cpumask)); in flush_tlb_others()
394 cpumask_and(&cpumask, &cpumask, cpu_online_mask); in flush_tlb_others()
[all …]
/arch/tile/include/asm/
Dsmp.h32 void send_IPI_many(const struct cpumask *mask, int tag);
74 static inline void arch_send_call_function_ipi_mask(struct cpumask *mask) in arch_send_call_function_ipi_mask()
101 extern struct cpumask cpu_lotar_map;
106 extern struct cpumask hash_for_home_map;
110 extern struct cpumask cpu_cacheable_map;
128 static inline int __cpulist_parse_crop(const char *buf, struct cpumask *dstp, in __cpulist_parse_crop()
Dhomecache.h60 const struct cpumask *cache_cpumask,
63 const struct cpumask *tlb_cpumask,
70 extern void homecache_evict(const struct cpumask *mask);
/arch/x86/include/asm/
Dapic.h296 const struct cpumask *(*target_cpus)(void);
304 void (*vector_allocation_domain)(int cpu, struct cpumask *retmask);
329 unsigned int (*cpu_mask_to_apicid)(const struct cpumask *cpumask);
330 unsigned int (*cpu_mask_to_apicid_and)(const struct cpumask *cpumask,
331 const struct cpumask *andmask);
334 void (*send_IPI_mask)(const struct cpumask *mask, int vector);
335 void (*send_IPI_mask_allbutself)(const struct cpumask *mask,
514 static inline const struct cpumask *default_target_cpus(void) in default_target_cpus()
573 default_cpu_mask_to_apicid(const struct cpumask *cpumask) in default_cpu_mask_to_apicid() argument
575 return cpumask_bits(cpumask)[0] & APIC_ALL_CPUS; in default_cpu_mask_to_apicid()
[all …]
Dipi.h122 extern void default_send_IPI_mask_sequence_phys(const struct cpumask *mask,
124 extern void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask,
149 extern void default_send_IPI_mask_sequence_logical(const struct cpumask *mask,
151 extern void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask,
153 extern void default_send_IPI_mask_logical(const struct cpumask *mask,
Dsmp.h41 static inline struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask()
46 static inline struct cpumask *cpu_core_mask(int cpu) in cpu_core_mask()
51 static inline struct cpumask *cpu_llc_shared_mask(int cpu) in cpu_llc_shared_mask()
78 void (*send_call_func_ipi)(const struct cpumask *mask);
146 static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
163 void native_send_call_func_ipi(const struct cpumask *mask);
Dtlbflush.h116 static inline void native_flush_tlb_others(const struct cpumask *cpumask, in native_flush_tlb_others() argument
145 void native_flush_tlb_others(const struct cpumask *cpumask,
/arch/mn10300/mm/
Dtlb-smp.c52 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm,
94 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, in flush_tlb_others() argument
105 BUG_ON(cpumask_empty(&cpumask)); in flush_tlb_others()
106 BUG_ON(cpumask_test_cpu(smp_processor_id(), &cpumask)); in flush_tlb_others()
108 cpumask_and(&tmp, &cpumask, cpu_online_mask); in flush_tlb_others()
109 BUG_ON(!cpumask_equal(&cpumask, &tmp)); in flush_tlb_others()
122 atomic_set_mask(cpumask.bits[0], &flush_cpumask.bits[0]); in flush_tlb_others()
/arch/hexagon/include/asm/
Dsmp.h37 extern void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg);
40 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
/arch/tile/mm/
Dmigrate.h30 const unsigned long *cpumask);
45 const struct cpumask *cache_cpumask,
46 const struct cpumask *tlb_cpumask,
Dhomecache.c76 static void hv_flush_update(const struct cpumask *cache_cpumask, in hv_flush_update()
77 struct cpumask *tlb_cpumask, in hv_flush_update()
81 struct cpumask mask; in hv_flush_update()
121 const struct cpumask *cache_cpumask_orig, in flush_remote()
124 const struct cpumask *tlb_cpumask_orig, in flush_remote()
129 struct cpumask cache_cpumask_copy, tlb_cpumask_copy; in flush_remote()
130 struct cpumask *cache_cpumask, *tlb_cpumask; in flush_remote()
200 void homecache_evict(const struct cpumask *mask) in homecache_evict()
211 struct cpumask *home_mask) in homecache_mask()
254 struct cpumask home_mask; in homecache_flush_cache()
/arch/x86/mm/
Dtlb.c172 static void flush_tlb_others_ipi(const struct cpumask *cpumask, in flush_tlb_others_ipi() argument
187 if (cpumask_andnot(to_cpumask(f->flush_cpumask), cpumask, cpumask_of(smp_processor_id()))) { in flush_tlb_others_ipi()
205 void native_flush_tlb_others(const struct cpumask *cpumask, in native_flush_tlb_others() argument
212 cpumask = uv_flush_tlb_others(cpumask, mm, va, cpu); in native_flush_tlb_others()
213 if (cpumask) in native_flush_tlb_others()
214 flush_tlb_others_ipi(cpumask, mm, va); in native_flush_tlb_others()
217 flush_tlb_others_ipi(cpumask, mm, va); in native_flush_tlb_others()
/arch/ia64/include/asm/
Dirq.h30 bool is_affinity_mask_valid(const struct cpumask *cpumask);
/arch/ia64/kernel/
Dsmp.c169 send_IPI_mask(const struct cpumask *mask, int op) in send_IPI_mask()
259 cpumask_t cpumask = xcpumask; in smp_flush_tlb_cpumask() local
265 for_each_cpu_mask(cpu, cpumask) in smp_flush_tlb_cpumask()
269 for_each_cpu_mask(cpu, cpumask) { in smp_flush_tlb_cpumask()
279 for_each_cpu_mask(cpu, cpumask) in smp_flush_tlb_cpumask()
324 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
/arch/hexagon/kernel/
Dsmp.c109 void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg) in send_ipi() argument
117 for_each_cpu(cpu, cpumask) { in send_ipi()
252 struct cpumask targets; in smp_send_stop()
263 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()

123456789