Searched refs:affinity (Results 1 – 11 of 11) sorted by relevance
/kernel/irq/ |
D | irqdesc.c | 57 if (!zalloc_cpumask_var_node(&desc->irq_common_data.affinity, in alloc_masks() 64 free_cpumask_var(desc->irq_common_data.affinity); in alloc_masks() 74 free_cpumask_var(desc->irq_common_data.affinity); in alloc_masks() 82 const struct cpumask *affinity) in desc_smp_init() argument 84 if (!affinity) in desc_smp_init() 85 affinity = irq_default_affinity; in desc_smp_init() 86 cpumask_copy(desc->irq_common_data.affinity, affinity); in desc_smp_init() 100 desc_smp_init(struct irq_desc *desc, int node, const struct cpumask *affinity) { } in desc_smp_init() argument 104 const struct cpumask *affinity, struct module *owner) in desc_set_defaults() argument 127 desc_smp_init(desc, node, affinity); in desc_set_defaults() [all …]
|
D | cpuhotplug.c | 58 const struct cpumask *affinity; in migrate_one_irq() local 105 affinity = irq_desc_get_pending_mask(desc); in migrate_one_irq() 107 affinity = irq_data_get_affinity_mask(d); in migrate_one_irq() 113 if (cpumask_any_and(affinity, cpu_online_mask) >= nr_cpu_ids) { in migrate_one_irq() 123 affinity = cpu_online_mask; in migrate_one_irq() 132 err = irq_do_set_affinity(d, affinity, false); in migrate_one_irq() 192 const struct cpumask *affinity = irq_data_get_affinity_mask(data); in irq_restore_affinity_of_irq() local 195 !irq_data_get_irq_chip(data) || !cpumask_test_cpu(cpu, affinity)) in irq_restore_affinity_of_irq() 211 irq_set_affinity_locked(data, affinity, false); in irq_restore_affinity_of_irq()
|
D | irqdomain.c | 30 bool realloc, const struct irq_affinity_desc *affinity); 711 const struct irq_affinity_desc *affinity) in irq_create_mapping_affinity_locked() argument 720 affinity); in irq_create_mapping_affinity_locked() 750 const struct irq_affinity_desc *affinity) in irq_create_mapping_affinity() argument 771 virq = irq_create_mapping_affinity_locked(domain, hwirq, affinity); in irq_create_mapping_affinity() 1102 int node, const struct irq_affinity_desc *affinity) in irq_domain_alloc_descs() argument 1108 affinity); in irq_domain_alloc_descs() 1114 affinity); in irq_domain_alloc_descs() 1117 affinity); in irq_domain_alloc_descs() 1491 bool realloc, const struct irq_affinity_desc *affinity) in irq_domain_alloc_irqs_locked() argument [all …]
|
D | msi.c | 36 const struct irq_affinity_desc *affinity) in msi_alloc_desc() argument 45 if (affinity) { in msi_alloc_desc() 46 desc->affinity = kmemdup(affinity, nvec * sizeof(*desc->affinity), GFP_KERNEL); in msi_alloc_desc() 47 if (!desc->affinity) { in msi_alloc_desc() 57 kfree(desc->affinity); in msi_free_desc() 85 desc = msi_alloc_desc(dev, init_desc->nvec_used, init_desc->affinity); in msi_add_msi_desc() 885 desc->affinity); in __msi_domain_alloc_irqs()
|
D | devres.c | 180 const struct irq_affinity_desc *affinity) in __devm_irq_alloc_descs() argument 189 base = __irq_alloc_descs(irq, from, cnt, node, owner, affinity); in __devm_irq_alloc_descs()
|
D | manage.c | 279 cpumask_copy(desc->irq_common_data.affinity, mask); in irq_do_set_affinity() 342 cpumask_copy(desc->irq_common_data.affinity, mask); in irq_set_affinity_deactivated() 397 struct irq_affinity_desc *affinity) in irq_update_affinity_desc() argument 435 if (affinity->is_managed) { in irq_update_affinity_desc() 440 cpumask_copy(desc->irq_common_data.affinity, &affinity->mask); in irq_update_affinity_desc() 528 cpumask_copy(cpumask, desc->irq_common_data.affinity); in irq_affinity_notify() 608 if (cpumask_intersects(desc->irq_common_data.affinity, in irq_setup_affinity() 610 set = desc->irq_common_data.affinity; in irq_setup_affinity() 1152 if (cpumask_available(desc->irq_common_data.affinity)) { in irq_thread_check_affinity()
|
D | Kconfig | 25 # Supports effective affinity mask 38 # Alpha specific irq affinity mechanism
|
D | Makefile | 18 obj-$(CONFIG_SMP) += affinity.o
|
D | ipi.c | 94 cpumask_copy(data->common->affinity, dest); in irq_reserve_ipi()
|
D | proc.c | 54 mask = desc->irq_common_data.affinity; in show_irq_affinity()
|
/kernel/rcu/ |
D | Kconfig | 260 preempted between each callback, and (2) affinity or cgroups can
|