Searched refs:cpumask_of_node (Results 1 – 25 of 27) sorted by relevance
12
/drivers/nvdimm/ |
D | nd_perf.c | 157 cpumask = cpumask_of_node(nodeid); in nvdimm_pmu_cpu_offline() 233 cpumask = cpumask_of_node(nodeid); in nvdimm_pmu_cpu_hotplug_init()
|
/drivers/base/ |
D | arch_numa.c | 46 const struct cpumask *cpumask_of_node(int node) in cpumask_of_node() function 60 EXPORT_SYMBOL(cpumask_of_node);
|
D | node.c | 42 cpumask_and(mask, cpumask_of_node(node_dev->dev.id), cpu_online_mask); in cpumap_read() 63 cpumask_and(mask, cpumask_of_node(node_dev->dev.id), cpu_online_mask); in cpulist_read()
|
D | arch_topology.c | 714 const cpumask_t *core_mask = cpumask_of_node(cpu_to_node(cpu)); in cpu_coregroup_mask()
|
/drivers/hwtracing/ptt/ |
D | hisi_ptt.c | 490 const cpumask_t *cpumask = cpumask_of_node(dev_to_node(&hisi_ptt->pdev->dev)); in cpumask_show() 824 if (!cpumask_test_cpu(cpu, cpumask_of_node(dev_to_node(&hisi_ptt->pdev->dev)))) in hisi_ptt_pmu_add() 1016 target = cpumask_any_but(cpumask_of_node(dev_to_node(&hisi_ptt->pdev->dev)), cpu); in hisi_ptt_cpu_teardown()
|
/drivers/md/ |
D | dm-ps-io-affinity.c | 219 cpumask = cpumask_of_node(node); in ioa_select_path()
|
/drivers/infiniband/hw/hfi1/ |
D | netdev_rx.c | 176 cpumask_and(node_cpu_mask, cpu_mask, cpumask_of_node(dd->node)); in hfi1_num_netdev_contexts()
|
D | affinity.c | 600 local_mask = cpumask_of_node(dd->node); in hfi1_dev_affinity_init() 1112 node_mask = cpumask_of_node(node); in hfi1_get_proc_affinity()
|
D | ruc.c | 548 cpumask_first(cpumask_of_node(ps.ppd->dd->node)); in hfi1_do_send()
|
D | qp.c | 337 cpumask_first(cpumask_of_node(dd->node))); in _hfi1_schedule_send()
|
D | tid_rdma.c | 629 cpumask_first(cpumask_of_node(dd->node)), in tid_rdma_schedule_tid_wakeup() 5376 cpumask_first(cpumask_of_node(ps.ppd->dd->node)); in hfi1_do_tid_send() 5430 cpumask_first(cpumask_of_node(dd->node))); in _hfi1_schedule_tid_send()
|
/drivers/perf/ |
D | thunderx2_pmu.c | 753 cpu = cpumask_any_and(cpumask_of_node(tx2_pmu->node), in tx2_uncore_pmu_add_dev() 951 cpumask_of_node(tx2_pmu->node), in tx2_uncore_pmu_offline_cpu()
|
D | alibaba_uncore_drw_pmu.c | 733 cpumask_of_node(cpu_to_node(cpu)), cpu_online_mask); in ali_drw_pmu_offline_cpu()
|
D | arm-cmn.c | 1835 if (cpumask_and(&mask, cpumask_of_node(node), cpu_online_mask) && in arm_cmn_pmu_offline_cpu()
|
/drivers/hv/ |
D | channel_mgmt.c | 765 if (cpumask_empty(cpumask_of_node(numa_node))) in init_vp_index() 772 cpumask_xor(available_mask, allocated_mask, cpumask_of_node(numa_node)); in init_vp_index()
|
/drivers/infiniband/sw/rdmavt/ |
D | cq.c | 247 cpumask_first(cpumask_of_node(rdi->dparms.node)); in rvt_create_cq()
|
/drivers/scsi/ |
D | storvsc_drv.c | 1396 node_mask = cpumask_of_node(cpu_to_node(q_num)); in get_og_chn() 1456 node_mask = cpumask_of_node(cpu_to_node(q_num)); in storvsc_do_io()
|
/drivers/misc/sgi-xp/ |
D | xpc_uv.c | 1747 for_each_cpu(cpu, cpumask_of_node(nid)) { in xpc_init_mq_node() 1760 for_each_cpu(cpu, cpumask_of_node(nid)) { in xpc_init_mq_node()
|
/drivers/pci/ |
D | pci-driver.c | 384 cpu = cpumask_any_and(cpumask_of_node(node), in pci_call_probe()
|
D | pci-sysfs.c | 107 mask = cpumask_of_node(dev_to_node(dev)); in pci_dev_show_local_cpu()
|
/drivers/dma/ |
D | dmaengine.c | 306 cpumask_test_cpu(cpu, cpumask_of_node(node)); in dma_chan_is_local()
|
/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_crat.c | 1642 cpumask = cpumask_of_node(numa_node_id); in kfd_fill_cu_for_cpu()
|
D | kfd_topology.c | 2170 return kfd_cpumask_to_apic_id(cpumask_of_node(numa_node_id)); in kfd_numa_node_to_apic_id()
|
/drivers/block/mtip32xx/ |
D | mtip32xx.c | 3598 node_mask = cpumask_of_node(node); in get_least_used_cpu_on_node() 3760 node_mask = cpumask_of_node(dd->numa_node); in mtip_pci_probe()
|
/drivers/irqchip/ |
D | irq-gic-v3-its.c | 1619 cpumask_and(tmpmask, cpumask_of_node(node), aff_mask); in its_select_cpu() 1659 cpumask_and(tmpmask, tmpmask, cpumask_of_node(node)); in its_select_cpu()
|
12