/arch/x86/include/asm/uv/ |
D | uv.h | 6 struct cpumask; 16 extern const struct cpumask *uv_flush_tlb_others(const struct cpumask *cpumask, 27 static inline const struct cpumask * 28 uv_flush_tlb_others(const struct cpumask *cpumask, struct mm_struct *mm, in uv_flush_tlb_others() argument 30 { return cpumask; } in uv_flush_tlb_others()
|
/arch/x86/kernel/apic/ |
D | apic_flat_64.c | 39 static const struct cpumask *flat_target_cpus(void) in flat_target_cpus() 44 static void flat_vector_allocation_domain(int cpu, struct cpumask *retmask) in flat_vector_allocation_domain() 87 static void flat_send_IPI_mask(const struct cpumask *cpumask, int vector) in flat_send_IPI_mask() argument 89 unsigned long mask = cpumask_bits(cpumask)[0]; in flat_send_IPI_mask() 95 flat_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) in flat_send_IPI_mask_allbutself() argument 97 unsigned long mask = cpumask_bits(cpumask)[0]; in flat_send_IPI_mask_allbutself() 264 static const struct cpumask *physflat_target_cpus(void) in physflat_target_cpus() 269 static void physflat_vector_allocation_domain(int cpu, struct cpumask *retmask) in physflat_vector_allocation_domain() 275 static void physflat_send_IPI_mask(const struct cpumask *cpumask, int vector) in physflat_send_IPI_mask() argument 277 default_send_IPI_mask_sequence_phys(cpumask, vector); in physflat_send_IPI_mask() [all …]
|
D | x2apic_phys.c | 39 __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) in __x2apic_send_IPI_mask() 59 static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask() 65 x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask_allbutself() 80 static unsigned int x2apic_cpu_mask_to_apicid(const struct cpumask *cpumask) in x2apic_cpu_mask_to_apicid() argument 86 int cpu = cpumask_first(cpumask); in x2apic_cpu_mask_to_apicid() 95 x2apic_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in x2apic_cpu_mask_to_apicid_and() argument 96 const struct cpumask *andmask) in x2apic_cpu_mask_to_apicid_and() 104 for_each_cpu_and(cpu, cpumask, andmask) { in x2apic_cpu_mask_to_apicid_and()
|
D | ipi.c | 22 void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector) in default_send_IPI_mask_sequence_phys() 40 void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, in default_send_IPI_mask_allbutself_phys() 61 void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, in default_send_IPI_mask_sequence_logical() 81 void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, in default_send_IPI_mask_allbutself_logical() 104 void default_send_IPI_mask_logical(const struct cpumask *cpumask, int vector) in default_send_IPI_mask_logical() argument 106 unsigned long mask = cpumask_bits(cpumask)[0]; in default_send_IPI_mask_logical()
|
D | bigsmp_32.c | 29 static const struct cpumask *bigsmp_target_cpus(void) in bigsmp_target_cpus() 109 static unsigned int bigsmp_cpu_mask_to_apicid(const struct cpumask *cpumask) in bigsmp_cpu_mask_to_apicid() argument 111 int cpu = cpumask_first(cpumask); in bigsmp_cpu_mask_to_apicid() 118 static unsigned int bigsmp_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in bigsmp_cpu_mask_to_apicid_and() argument 119 const struct cpumask *andmask) in bigsmp_cpu_mask_to_apicid_and() 127 for_each_cpu_and(cpu, cpumask, andmask) { in bigsmp_cpu_mask_to_apicid_and() 139 static inline void bigsmp_send_IPI_mask(const struct cpumask *mask, int vector) in bigsmp_send_IPI_mask() 180 static void bigsmp_vector_allocation_domain(int cpu, struct cpumask *retmask) in bigsmp_vector_allocation_domain()
|
D | x2apic_cluster.c | 28 __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) in __x2apic_send_IPI_mask() 30 struct cpumask *cpus_in_cluster_ptr; in __x2apic_send_IPI_mask() 31 struct cpumask *ipi_mask_ptr; in __x2apic_send_IPI_mask() 78 static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask() 84 x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) in x2apic_send_IPI_mask_allbutself() 99 static unsigned int x2apic_cpu_mask_to_apicid(const struct cpumask *cpumask) in x2apic_cpu_mask_to_apicid() argument 105 int cpu = cpumask_first(cpumask); in x2apic_cpu_mask_to_apicid() 114 x2apic_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in x2apic_cpu_mask_to_apicid_and() argument 115 const struct cpumask *andmask) in x2apic_cpu_mask_to_apicid_and() 123 for_each_cpu_and(cpu, cpumask, andmask) { in x2apic_cpu_mask_to_apicid_and()
|
D | apic_numachip.c | 75 static const struct cpumask *numachip_target_cpus(void) in numachip_target_cpus() 80 static void numachip_vector_allocation_domain(int cpu, struct cpumask *retmask) in numachip_vector_allocation_domain() 119 static void numachip_send_IPI_mask(const struct cpumask *mask, int vector) in numachip_send_IPI_mask() 127 static void numachip_send_IPI_mask_allbutself(const struct cpumask *mask, in numachip_send_IPI_mask_allbutself() 160 static unsigned int numachip_cpu_mask_to_apicid(const struct cpumask *cpumask) in numachip_cpu_mask_to_apicid() argument 168 cpu = cpumask_first(cpumask); in numachip_cpu_mask_to_apicid() 176 numachip_cpu_mask_to_apicid_and(const struct cpumask *cpumask, in numachip_cpu_mask_to_apicid_and() argument 177 const struct cpumask *andmask) in numachip_cpu_mask_to_apicid_and() 185 for_each_cpu_and(cpu, cpumask, andmask) { in numachip_cpu_mask_to_apicid_and()
|
D | summit_32.c | 55 static inline void summit_send_IPI_mask(const struct cpumask *mask, int vector) in summit_send_IPI_mask() 177 static const struct cpumask *summit_target_cpus(void) in summit_target_cpus() 266 static unsigned int summit_cpu_mask_to_apicid(const struct cpumask *cpumask) in summit_cpu_mask_to_apicid() argument 274 for_each_cpu(cpu, cpumask) { in summit_cpu_mask_to_apicid() 287 static unsigned int summit_cpu_mask_to_apicid_and(const struct cpumask *inmask, in summit_cpu_mask_to_apicid_and() 288 const struct cpumask *andmask) in summit_cpu_mask_to_apicid_and() 291 cpumask_var_t cpumask; in summit_cpu_mask_to_apicid_and() local 293 if (!alloc_cpumask_var(&cpumask, GFP_ATOMIC)) in summit_cpu_mask_to_apicid_and() 296 cpumask_and(cpumask, inmask, andmask); in summit_cpu_mask_to_apicid_and() 297 cpumask_and(cpumask, cpumask, cpu_online_mask); in summit_cpu_mask_to_apicid_and() [all …]
|
D | apic_noop.c | 34 static void noop_send_IPI_mask(const struct cpumask *cpumask, int vector) { } in noop_send_IPI_mask() argument 35 static void noop_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) { } in noop_send_IPI_mask_allbutself() argument 87 static const struct cpumask *noop_target_cpus(void) in noop_target_cpus() 103 static void noop_vector_allocation_domain(int cpu, struct cpumask *retmask) in noop_vector_allocation_domain()
|
D | es7000_32.c | 397 static void es7000_vector_allocation_domain(int cpu, struct cpumask *retmask) in es7000_vector_allocation_domain() 423 static void es7000_send_IPI_mask(const struct cpumask *mask, int vector) in es7000_send_IPI_mask() 443 static const struct cpumask *target_cpus_cluster(void) in target_cpus_cluster() 448 static const struct cpumask *es7000_target_cpus(void) in es7000_target_cpus() 543 static unsigned int es7000_cpu_mask_to_apicid(const struct cpumask *cpumask) in es7000_cpu_mask_to_apicid() argument 551 for_each_cpu(cpu, cpumask) { in es7000_cpu_mask_to_apicid() 566 es7000_cpu_mask_to_apicid_and(const struct cpumask *inmask, in es7000_cpu_mask_to_apicid_and() 567 const struct cpumask *andmask) in es7000_cpu_mask_to_apicid_and() 570 cpumask_var_t cpumask; in es7000_cpu_mask_to_apicid_and() local 572 if (!alloc_cpumask_var(&cpumask, GFP_ATOMIC)) in es7000_cpu_mask_to_apicid_and() [all …]
|
/arch/m32r/kernel/ |
D | smp.c | 80 static void send_IPI_mask(const struct cpumask *, int, int); 151 cpumask_t cpumask; in smp_flush_cache_all() local 155 cpumask_copy(&cpumask, cpu_online_mask); in smp_flush_cache_all() 156 cpumask_clear_cpu(smp_processor_id(), &cpumask); in smp_flush_cache_all() 158 mask=cpumask_bits(&cpumask); in smp_flush_cache_all() 160 send_IPI_mask(&cpumask, INVALIDATE_CACHE_IPI, 0); in smp_flush_cache_all() 370 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, in flush_tlb_others() argument 388 BUG_ON(cpumask_empty(&cpumask)); in flush_tlb_others() 390 BUG_ON(cpumask_test_cpu(smp_processor_id(), &cpumask)); in flush_tlb_others() 394 cpumask_and(&cpumask, &cpumask, cpu_online_mask); in flush_tlb_others() [all …]
|
/arch/tile/include/asm/ |
D | smp.h | 32 void send_IPI_many(const struct cpumask *mask, int tag); 74 static inline void arch_send_call_function_ipi_mask(struct cpumask *mask) in arch_send_call_function_ipi_mask() 101 extern struct cpumask cpu_lotar_map; 106 extern struct cpumask hash_for_home_map; 110 extern struct cpumask cpu_cacheable_map; 128 static inline int __cpulist_parse_crop(const char *buf, struct cpumask *dstp, in __cpulist_parse_crop()
|
D | homecache.h | 60 const struct cpumask *cache_cpumask, 63 const struct cpumask *tlb_cpumask, 70 extern void homecache_evict(const struct cpumask *mask);
|
/arch/x86/include/asm/ |
D | apic.h | 296 const struct cpumask *(*target_cpus)(void); 304 void (*vector_allocation_domain)(int cpu, struct cpumask *retmask); 329 unsigned int (*cpu_mask_to_apicid)(const struct cpumask *cpumask); 330 unsigned int (*cpu_mask_to_apicid_and)(const struct cpumask *cpumask, 331 const struct cpumask *andmask); 334 void (*send_IPI_mask)(const struct cpumask *mask, int vector); 335 void (*send_IPI_mask_allbutself)(const struct cpumask *mask, 514 static inline const struct cpumask *default_target_cpus(void) in default_target_cpus() 573 default_cpu_mask_to_apicid(const struct cpumask *cpumask) in default_cpu_mask_to_apicid() argument 575 return cpumask_bits(cpumask)[0] & APIC_ALL_CPUS; in default_cpu_mask_to_apicid() [all …]
|
D | ipi.h | 122 extern void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, 124 extern void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, 149 extern void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, 151 extern void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, 153 extern void default_send_IPI_mask_logical(const struct cpumask *mask,
|
D | smp.h | 41 static inline struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask() 46 static inline struct cpumask *cpu_core_mask(int cpu) in cpu_core_mask() 51 static inline struct cpumask *cpu_llc_shared_mask(int cpu) in cpu_llc_shared_mask() 78 void (*send_call_func_ipi)(const struct cpumask *mask); 146 static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask() 163 void native_send_call_func_ipi(const struct cpumask *mask);
|
D | tlbflush.h | 116 static inline void native_flush_tlb_others(const struct cpumask *cpumask, in native_flush_tlb_others() argument 145 void native_flush_tlb_others(const struct cpumask *cpumask,
|
/arch/mn10300/mm/ |
D | tlb-smp.c | 52 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, 94 static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, in flush_tlb_others() argument 105 BUG_ON(cpumask_empty(&cpumask)); in flush_tlb_others() 106 BUG_ON(cpumask_test_cpu(smp_processor_id(), &cpumask)); in flush_tlb_others() 108 cpumask_and(&tmp, &cpumask, cpu_online_mask); in flush_tlb_others() 109 BUG_ON(!cpumask_equal(&cpumask, &tmp)); in flush_tlb_others() 122 atomic_set_mask(cpumask.bits[0], &flush_cpumask.bits[0]); in flush_tlb_others()
|
/arch/hexagon/include/asm/ |
D | smp.h | 37 extern void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg); 40 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
/arch/tile/mm/ |
D | migrate.h | 30 const unsigned long *cpumask); 45 const struct cpumask *cache_cpumask, 46 const struct cpumask *tlb_cpumask,
|
D | homecache.c | 76 static void hv_flush_update(const struct cpumask *cache_cpumask, in hv_flush_update() 77 struct cpumask *tlb_cpumask, in hv_flush_update() 81 struct cpumask mask; in hv_flush_update() 121 const struct cpumask *cache_cpumask_orig, in flush_remote() 124 const struct cpumask *tlb_cpumask_orig, in flush_remote() 129 struct cpumask cache_cpumask_copy, tlb_cpumask_copy; in flush_remote() 130 struct cpumask *cache_cpumask, *tlb_cpumask; in flush_remote() 200 void homecache_evict(const struct cpumask *mask) in homecache_evict() 211 struct cpumask *home_mask) in homecache_mask() 254 struct cpumask home_mask; in homecache_flush_cache()
|
/arch/x86/mm/ |
D | tlb.c | 172 static void flush_tlb_others_ipi(const struct cpumask *cpumask, in flush_tlb_others_ipi() argument 187 if (cpumask_andnot(to_cpumask(f->flush_cpumask), cpumask, cpumask_of(smp_processor_id()))) { in flush_tlb_others_ipi() 205 void native_flush_tlb_others(const struct cpumask *cpumask, in native_flush_tlb_others() argument 212 cpumask = uv_flush_tlb_others(cpumask, mm, va, cpu); in native_flush_tlb_others() 213 if (cpumask) in native_flush_tlb_others() 214 flush_tlb_others_ipi(cpumask, mm, va); in native_flush_tlb_others() 217 flush_tlb_others_ipi(cpumask, mm, va); in native_flush_tlb_others()
|
/arch/ia64/include/asm/ |
D | irq.h | 30 bool is_affinity_mask_valid(const struct cpumask *cpumask);
|
/arch/ia64/kernel/ |
D | smp.c | 169 send_IPI_mask(const struct cpumask *mask, int op) in send_IPI_mask() 259 cpumask_t cpumask = xcpumask; in smp_flush_tlb_cpumask() local 265 for_each_cpu_mask(cpu, cpumask) in smp_flush_tlb_cpumask() 269 for_each_cpu_mask(cpu, cpumask) { in smp_flush_tlb_cpumask() 279 for_each_cpu_mask(cpu, cpumask) in smp_flush_tlb_cpumask() 324 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
|
/arch/hexagon/kernel/ |
D | smp.c | 109 void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg) in send_ipi() argument 117 for_each_cpu(cpu, cpumask) { in send_ipi() 252 struct cpumask targets; in smp_send_stop() 263 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
|