/drivers/base/ |
D | node.c | 63 int nid = dev->id; in node_read_meminfo() local 66 si_meminfo_node(&i, nid); in node_read_meminfo() 79 nid, K(i.totalram), in node_read_meminfo() 80 nid, K(i.freeram), in node_read_meminfo() 81 nid, K(i.totalram - i.freeram), in node_read_meminfo() 82 nid, K(node_page_state(nid, NR_ACTIVE_ANON) + in node_read_meminfo() 83 node_page_state(nid, NR_ACTIVE_FILE)), in node_read_meminfo() 84 nid, K(node_page_state(nid, NR_INACTIVE_ANON) + in node_read_meminfo() 85 node_page_state(nid, NR_INACTIVE_FILE)), in node_read_meminfo() 86 nid, K(node_page_state(nid, NR_ACTIVE_ANON)), in node_read_meminfo() [all …]
|
D | memory.c | 398 int nid; in memory_probe_store() local 408 nid = memory_add_physaddr_to_nid(phys_addr); in memory_probe_store() 409 ret = add_memory(nid, phys_addr, in memory_probe_store() 576 static int add_memory_section(int nid, struct mem_section *section, in add_memory_section() argument 611 ret = register_mem_sect_under_node(mem, nid); in add_memory_section() 647 int register_new_memory(int nid, struct mem_section *section) in register_new_memory() argument 649 return add_memory_section(nid, section, NULL, MEM_OFFLINE, HOTPLUG); in register_new_memory()
|
/drivers/char/ |
D | mspec.c | 115 int nid; in mspec_zero_block() local 119 nid = nasid_to_cnodeid(get_node_number(__pa(addr))); in mspec_zero_block() 120 p = (void *)TO_AMO(scratch_page[nid]); in mspec_zero_block() 360 int nid; in mspec_init() local 371 for_each_node_state(nid, N_ONLINE) { in mspec_init() 376 scratch_page[nid] = uncached_alloc_page(nid, 1); in mspec_init() 377 if (scratch_page[nid] == 0) in mspec_init() 379 phys = __pa(scratch_page[nid]); in mspec_init() 382 if (actual_nid != nid) in mspec_init() 421 for_each_node(nid) { in mspec_init() [all …]
|
/drivers/edac/ |
D | amd64_edac.c | 242 unsigned nid) in amd64_base_limit_match() argument 254 return ((addr >= get_dram_base(pvt, nid)) && in amd64_base_limit_match() 255 (addr <= get_dram_limit(pvt, nid))); in amd64_base_limit_match() 1027 u8 nid = dram_dst_node(pvt, range); in read_dram_base_limit_regs() local 1030 f1 = pci_get_domain_bus_and_slot(0, 0, PCI_DEVFN(0x18 + nid, 1)); in read_dram_base_limit_regs() 1414 static int f1x_lookup_addr_in_dct(u64 in_addr, u32 nid, u8 dct) in f1x_lookup_addr_in_dct() argument 1422 mci = mcis[nid]; in f1x_lookup_addr_in_dct() 1493 u64 sys_addr, int *nid, int *chan_sel) in f1x_match_to_this_node() argument 1567 *nid = node_id; in f1x_match_to_this_node() 1609 int nid, csrow, chan = 0; in f1x_map_sysaddr_to_csrow() local [all …]
|
/drivers/char/agp/ |
D | sgi-agp.c | 45 int nid; in sgi_tioca_alloc_page() local 49 nid = info->ca_closest_node; in sgi_tioca_alloc_page() 50 page = alloc_pages_node(nid, GFP_KERNEL, 0); in sgi_tioca_alloc_page()
|
/drivers/misc/sgi-xp/ |
D | xpc_uv.c | 211 int nid; in xpc_create_gru_mq_uv() local 241 nid = cpu_to_node(cpu); in xpc_create_gru_mq_uv() 242 page = alloc_pages_exact_node(nid, in xpc_create_gru_mq_uv() 247 "bytes of memory on nid=%d for GRU mq\n", mq_size, nid); in xpc_create_gru_mq_uv() 1737 xpc_init_mq_node(int nid) in xpc_init_mq_node() argument 1743 for_each_cpu(cpu, cpumask_of_node(nid)) { in xpc_init_mq_node() 1745 xpc_create_gru_mq_uv(XPC_ACTIVATE_MQ_SIZE_UV, nid, in xpc_init_mq_node() 1756 for_each_cpu(cpu, cpumask_of_node(nid)) { in xpc_init_mq_node() 1758 xpc_create_gru_mq_uv(XPC_NOTIFY_MQ_SIZE_UV, nid, in xpc_init_mq_node() 1777 int nid; in xpc_init_uv() local [all …]
|
/drivers/xen/ |
D | balloon.c | 235 int nid, rc; in reserve_additional_memory() local 241 nid = memory_add_physaddr_to_nid(hotplug_start_paddr); in reserve_additional_memory() 243 rc = add_memory(nid, hotplug_start_paddr, balloon_hotplug << PAGE_SHIFT); in reserve_additional_memory()
|
/drivers/pcmcia/ |
D | cistpl.c | 825 int nid; in parse_jedec() local 830 for (nid = 0; nid < CISTPL_MAX_DEVICES; nid++) { in parse_jedec() 833 jedec->id[nid].mfr = p[0]; in parse_jedec() 834 jedec->id[nid].info = p[1]; in parse_jedec() 837 jedec->nid = nid; in parse_jedec()
|
/drivers/misc/sgi-gru/ |
D | grufile.c | 261 int pnode, nid, bid, chip; in gru_init_tables() local 273 nid = uv_blade_to_memory_nid(bid);/* -1 if no memory on blade */ in gru_init_tables() 274 page = alloc_pages_node(nid, GFP_KERNEL, order); in gru_init_tables()
|
D | grutables.h | 522 #define for_each_gru_on_blade(gru, nid, i) \ argument 523 for ((gru) = gru_base[nid]->bs_grus, (i) = 0; \
|
/drivers/iommu/ |
D | intel-iommu.c | 366 int nid; /* node id */ member 797 tmp_page = alloc_pgtable_page(domain->nid); in pfn_to_dma_pte() 1342 domain->nid = -1; in alloc_domain() 1502 domain->nid = iommu->node; in domain_init() 1505 domain->pgd = (struct dma_pte *)alloc_pgtable_page(domain->nid); in domain_init() 1665 domain->nid = iommu->node; in domain_context_mapping_one() 2227 int nid, ret = 0; in si_domain_init() local 2253 for_each_online_node(nid) { in si_domain_init() 2257 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in si_domain_init() 3872 domain->nid = -1; in iommu_alloc_vm_domain() [all …]
|
/drivers/mtd/maps/ |
D | pcmciamtd.c | 358 for (i = 0; i < t->nid; i++) in pcmciamtd_cistpl_jedec()
|