• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * pseries Memory Hotplug infrastructure.
4  *
5  * Copyright (C) 2008 Badari Pulavarty, IBM Corporation
6  */
7 
8 #define pr_fmt(fmt)	"pseries-hotplug-mem: " fmt
9 
10 #include <linux/of.h>
11 #include <linux/of_address.h>
12 #include <linux/memblock.h>
13 #include <linux/memory.h>
14 #include <linux/memory_hotplug.h>
15 #include <linux/slab.h>
16 
17 #include <asm/firmware.h>
18 #include <asm/machdep.h>
19 #include <asm/prom.h>
20 #include <asm/sparsemem.h>
21 #include <asm/fadump.h>
22 #include <asm/drmem.h>
23 #include "pseries.h"
24 
25 static bool rtas_hp_event;
26 
pseries_memory_block_size(void)27 unsigned long pseries_memory_block_size(void)
28 {
29 	struct device_node *np;
30 	u64 memblock_size = MIN_MEMORY_BLOCK_SIZE;
31 	struct resource r;
32 
33 	np = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory");
34 	if (np) {
35 		const __be64 *size;
36 
37 		size = of_get_property(np, "ibm,lmb-size", NULL);
38 		if (size)
39 			memblock_size = be64_to_cpup(size);
40 		of_node_put(np);
41 	} else  if (machine_is(pseries)) {
42 		/* This fallback really only applies to pseries */
43 		unsigned int memzero_size = 0;
44 
45 		np = of_find_node_by_path("/memory@0");
46 		if (np) {
47 			if (!of_address_to_resource(np, 0, &r))
48 				memzero_size = resource_size(&r);
49 			of_node_put(np);
50 		}
51 
52 		if (memzero_size) {
53 			/* We now know the size of memory@0, use this to find
54 			 * the first memoryblock and get its size.
55 			 */
56 			char buf[64];
57 
58 			sprintf(buf, "/memory@%x", memzero_size);
59 			np = of_find_node_by_path(buf);
60 			if (np) {
61 				if (!of_address_to_resource(np, 0, &r))
62 					memblock_size = resource_size(&r);
63 				of_node_put(np);
64 			}
65 		}
66 	}
67 	return memblock_size;
68 }
69 
dlpar_free_property(struct property * prop)70 static void dlpar_free_property(struct property *prop)
71 {
72 	kfree(prop->name);
73 	kfree(prop->value);
74 	kfree(prop);
75 }
76 
dlpar_clone_property(struct property * prop,u32 prop_size)77 static struct property *dlpar_clone_property(struct property *prop,
78 					     u32 prop_size)
79 {
80 	struct property *new_prop;
81 
82 	new_prop = kzalloc(sizeof(*new_prop), GFP_KERNEL);
83 	if (!new_prop)
84 		return NULL;
85 
86 	new_prop->name = kstrdup(prop->name, GFP_KERNEL);
87 	new_prop->value = kzalloc(prop_size, GFP_KERNEL);
88 	if (!new_prop->name || !new_prop->value) {
89 		dlpar_free_property(new_prop);
90 		return NULL;
91 	}
92 
93 	memcpy(new_prop->value, prop->value, prop->length);
94 	new_prop->length = prop_size;
95 
96 	of_property_set_flag(new_prop, OF_DYNAMIC);
97 	return new_prop;
98 }
99 
find_aa_index(struct device_node * dr_node,struct property * ala_prop,const u32 * lmb_assoc,u32 * aa_index)100 static bool find_aa_index(struct device_node *dr_node,
101 			 struct property *ala_prop,
102 			 const u32 *lmb_assoc, u32 *aa_index)
103 {
104 	u32 *assoc_arrays, new_prop_size;
105 	struct property *new_prop;
106 	int aa_arrays, aa_array_entries, aa_array_sz;
107 	int i, index;
108 
109 	/*
110 	 * The ibm,associativity-lookup-arrays property is defined to be
111 	 * a 32-bit value specifying the number of associativity arrays
112 	 * followed by a 32-bitvalue specifying the number of entries per
113 	 * array, followed by the associativity arrays.
114 	 */
115 	assoc_arrays = ala_prop->value;
116 
117 	aa_arrays = be32_to_cpu(assoc_arrays[0]);
118 	aa_array_entries = be32_to_cpu(assoc_arrays[1]);
119 	aa_array_sz = aa_array_entries * sizeof(u32);
120 
121 	for (i = 0; i < aa_arrays; i++) {
122 		index = (i * aa_array_entries) + 2;
123 
124 		if (memcmp(&assoc_arrays[index], &lmb_assoc[1], aa_array_sz))
125 			continue;
126 
127 		*aa_index = i;
128 		return true;
129 	}
130 
131 	new_prop_size = ala_prop->length + aa_array_sz;
132 	new_prop = dlpar_clone_property(ala_prop, new_prop_size);
133 	if (!new_prop)
134 		return false;
135 
136 	assoc_arrays = new_prop->value;
137 
138 	/* increment the number of entries in the lookup array */
139 	assoc_arrays[0] = cpu_to_be32(aa_arrays + 1);
140 
141 	/* copy the new associativity into the lookup array */
142 	index = aa_arrays * aa_array_entries + 2;
143 	memcpy(&assoc_arrays[index], &lmb_assoc[1], aa_array_sz);
144 
145 	of_update_property(dr_node, new_prop);
146 
147 	/*
148 	 * The associativity lookup array index for this lmb is
149 	 * number of entries - 1 since we added its associativity
150 	 * to the end of the lookup array.
151 	 */
152 	*aa_index = be32_to_cpu(assoc_arrays[0]) - 1;
153 	return true;
154 }
155 
update_lmb_associativity_index(struct drmem_lmb * lmb)156 static int update_lmb_associativity_index(struct drmem_lmb *lmb)
157 {
158 	struct device_node *parent, *lmb_node, *dr_node;
159 	struct property *ala_prop;
160 	const u32 *lmb_assoc;
161 	u32 aa_index;
162 	bool found;
163 
164 	parent = of_find_node_by_path("/");
165 	if (!parent)
166 		return -ENODEV;
167 
168 	lmb_node = dlpar_configure_connector(cpu_to_be32(lmb->drc_index),
169 					     parent);
170 	of_node_put(parent);
171 	if (!lmb_node)
172 		return -EINVAL;
173 
174 	lmb_assoc = of_get_property(lmb_node, "ibm,associativity", NULL);
175 	if (!lmb_assoc) {
176 		dlpar_free_cc_nodes(lmb_node);
177 		return -ENODEV;
178 	}
179 
180 	dr_node = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory");
181 	if (!dr_node) {
182 		dlpar_free_cc_nodes(lmb_node);
183 		return -ENODEV;
184 	}
185 
186 	ala_prop = of_find_property(dr_node, "ibm,associativity-lookup-arrays",
187 				    NULL);
188 	if (!ala_prop) {
189 		of_node_put(dr_node);
190 		dlpar_free_cc_nodes(lmb_node);
191 		return -ENODEV;
192 	}
193 
194 	found = find_aa_index(dr_node, ala_prop, lmb_assoc, &aa_index);
195 
196 	of_node_put(dr_node);
197 	dlpar_free_cc_nodes(lmb_node);
198 
199 	if (!found) {
200 		pr_err("Could not find LMB associativity\n");
201 		return -1;
202 	}
203 
204 	lmb->aa_index = aa_index;
205 	return 0;
206 }
207 
lmb_to_memblock(struct drmem_lmb * lmb)208 static struct memory_block *lmb_to_memblock(struct drmem_lmb *lmb)
209 {
210 	unsigned long section_nr;
211 	struct mem_section *mem_sect;
212 	struct memory_block *mem_block;
213 
214 	section_nr = pfn_to_section_nr(PFN_DOWN(lmb->base_addr));
215 	mem_sect = __nr_to_section(section_nr);
216 
217 	mem_block = find_memory_block(mem_sect);
218 	return mem_block;
219 }
220 
get_lmb_range(u32 drc_index,int n_lmbs,struct drmem_lmb ** start_lmb,struct drmem_lmb ** end_lmb)221 static int get_lmb_range(u32 drc_index, int n_lmbs,
222 			 struct drmem_lmb **start_lmb,
223 			 struct drmem_lmb **end_lmb)
224 {
225 	struct drmem_lmb *lmb, *start, *end;
226 	struct drmem_lmb *limit;
227 
228 	start = NULL;
229 	for_each_drmem_lmb(lmb) {
230 		if (lmb->drc_index == drc_index) {
231 			start = lmb;
232 			break;
233 		}
234 	}
235 
236 	if (!start)
237 		return -EINVAL;
238 
239 	end = &start[n_lmbs];
240 
241 	limit = &drmem_info->lmbs[drmem_info->n_lmbs];
242 	if (end > limit)
243 		return -EINVAL;
244 
245 	*start_lmb = start;
246 	*end_lmb = end;
247 	return 0;
248 }
249 
dlpar_change_lmb_state(struct drmem_lmb * lmb,bool online)250 static int dlpar_change_lmb_state(struct drmem_lmb *lmb, bool online)
251 {
252 	struct memory_block *mem_block;
253 	int rc;
254 
255 	mem_block = lmb_to_memblock(lmb);
256 	if (!mem_block)
257 		return -EINVAL;
258 
259 	if (online && mem_block->dev.offline)
260 		rc = device_online(&mem_block->dev);
261 	else if (!online && !mem_block->dev.offline)
262 		rc = device_offline(&mem_block->dev);
263 	else
264 		rc = 0;
265 
266 	put_device(&mem_block->dev);
267 
268 	return rc;
269 }
270 
dlpar_online_lmb(struct drmem_lmb * lmb)271 static int dlpar_online_lmb(struct drmem_lmb *lmb)
272 {
273 	return dlpar_change_lmb_state(lmb, true);
274 }
275 
276 #ifdef CONFIG_MEMORY_HOTREMOVE
dlpar_offline_lmb(struct drmem_lmb * lmb)277 static int dlpar_offline_lmb(struct drmem_lmb *lmb)
278 {
279 	return dlpar_change_lmb_state(lmb, false);
280 }
281 
pseries_remove_memblock(unsigned long base,unsigned long memblock_size)282 static int pseries_remove_memblock(unsigned long base, unsigned long memblock_size)
283 {
284 	unsigned long block_sz, start_pfn;
285 	int sections_per_block;
286 	int i, nid;
287 
288 	start_pfn = base >> PAGE_SHIFT;
289 
290 	lock_device_hotplug();
291 
292 	if (!pfn_valid(start_pfn))
293 		goto out;
294 
295 	block_sz = pseries_memory_block_size();
296 	sections_per_block = block_sz / MIN_MEMORY_BLOCK_SIZE;
297 	nid = memory_add_physaddr_to_nid(base);
298 
299 	for (i = 0; i < sections_per_block; i++) {
300 		__remove_memory(nid, base, MIN_MEMORY_BLOCK_SIZE);
301 		base += MIN_MEMORY_BLOCK_SIZE;
302 	}
303 
304 out:
305 	/* Update memory regions for memory remove */
306 	memblock_remove(base, memblock_size);
307 	unlock_device_hotplug();
308 	return 0;
309 }
310 
pseries_remove_mem_node(struct device_node * np)311 static int pseries_remove_mem_node(struct device_node *np)
312 {
313 	const __be32 *prop;
314 	unsigned long base;
315 	unsigned long lmb_size;
316 	int ret = -EINVAL;
317 	int addr_cells, size_cells;
318 
319 	/*
320 	 * Check to see if we are actually removing memory
321 	 */
322 	if (!of_node_is_type(np, "memory"))
323 		return 0;
324 
325 	/*
326 	 * Find the base address and size of the memblock
327 	 */
328 	prop = of_get_property(np, "reg", NULL);
329 	if (!prop)
330 		return ret;
331 
332 	addr_cells = of_n_addr_cells(np);
333 	size_cells = of_n_size_cells(np);
334 
335 	/*
336 	 * "reg" property represents (addr,size) tuple.
337 	 */
338 	base = of_read_number(prop, addr_cells);
339 	prop += addr_cells;
340 	lmb_size = of_read_number(prop, size_cells);
341 
342 	pseries_remove_memblock(base, lmb_size);
343 	return 0;
344 }
345 
lmb_is_removable(struct drmem_lmb * lmb)346 static bool lmb_is_removable(struct drmem_lmb *lmb)
347 {
348 	int i, scns_per_block;
349 	int rc = 1;
350 	unsigned long pfn, block_sz;
351 	u64 phys_addr;
352 
353 	if (!(lmb->flags & DRCONF_MEM_ASSIGNED))
354 		return false;
355 
356 	block_sz = memory_block_size_bytes();
357 	scns_per_block = block_sz / MIN_MEMORY_BLOCK_SIZE;
358 	phys_addr = lmb->base_addr;
359 
360 #ifdef CONFIG_FA_DUMP
361 	/*
362 	 * Don't hot-remove memory that falls in fadump boot memory area
363 	 * and memory that is reserved for capturing old kernel memory.
364 	 */
365 	if (is_fadump_memory_area(phys_addr, block_sz))
366 		return false;
367 #endif
368 
369 	for (i = 0; i < scns_per_block; i++) {
370 		pfn = PFN_DOWN(phys_addr);
371 		if (!pfn_present(pfn)) {
372 			phys_addr += MIN_MEMORY_BLOCK_SIZE;
373 			continue;
374 		}
375 
376 		rc &= is_mem_section_removable(pfn, PAGES_PER_SECTION);
377 		phys_addr += MIN_MEMORY_BLOCK_SIZE;
378 	}
379 
380 	return rc ? true : false;
381 }
382 
383 static int dlpar_add_lmb(struct drmem_lmb *);
384 
dlpar_remove_lmb(struct drmem_lmb * lmb)385 static int dlpar_remove_lmb(struct drmem_lmb *lmb)
386 {
387 	struct memory_block *mem_block;
388 	unsigned long block_sz;
389 	int rc;
390 
391 	if (!lmb_is_removable(lmb))
392 		return -EINVAL;
393 
394 	mem_block = lmb_to_memblock(lmb);
395 	if (mem_block == NULL)
396 		return -EINVAL;
397 
398 	rc = dlpar_offline_lmb(lmb);
399 	if (rc) {
400 		put_device(&mem_block->dev);
401 		return rc;
402 	}
403 
404 	block_sz = pseries_memory_block_size();
405 
406 	__remove_memory(mem_block->nid, lmb->base_addr, block_sz);
407 	put_device(&mem_block->dev);
408 
409 	/* Update memory regions for memory remove */
410 	memblock_remove(lmb->base_addr, block_sz);
411 
412 	invalidate_lmb_associativity_index(lmb);
413 	lmb->flags &= ~DRCONF_MEM_ASSIGNED;
414 
415 	return 0;
416 }
417 
dlpar_memory_remove_by_count(u32 lmbs_to_remove)418 static int dlpar_memory_remove_by_count(u32 lmbs_to_remove)
419 {
420 	struct drmem_lmb *lmb;
421 	int lmbs_removed = 0;
422 	int lmbs_available = 0;
423 	int rc;
424 
425 	pr_info("Attempting to hot-remove %d LMB(s)\n", lmbs_to_remove);
426 
427 	if (lmbs_to_remove == 0)
428 		return -EINVAL;
429 
430 	/* Validate that there are enough LMBs to satisfy the request */
431 	for_each_drmem_lmb(lmb) {
432 		if (lmb_is_removable(lmb))
433 			lmbs_available++;
434 
435 		if (lmbs_available == lmbs_to_remove)
436 			break;
437 	}
438 
439 	if (lmbs_available < lmbs_to_remove) {
440 		pr_info("Not enough LMBs available (%d of %d) to satisfy request\n",
441 			lmbs_available, lmbs_to_remove);
442 		return -EINVAL;
443 	}
444 
445 	for_each_drmem_lmb(lmb) {
446 		rc = dlpar_remove_lmb(lmb);
447 		if (rc)
448 			continue;
449 
450 		/* Mark this lmb so we can add it later if all of the
451 		 * requested LMBs cannot be removed.
452 		 */
453 		drmem_mark_lmb_reserved(lmb);
454 
455 		lmbs_removed++;
456 		if (lmbs_removed == lmbs_to_remove)
457 			break;
458 	}
459 
460 	if (lmbs_removed != lmbs_to_remove) {
461 		pr_err("Memory hot-remove failed, adding LMB's back\n");
462 
463 		for_each_drmem_lmb(lmb) {
464 			if (!drmem_lmb_reserved(lmb))
465 				continue;
466 
467 			rc = dlpar_add_lmb(lmb);
468 			if (rc)
469 				pr_err("Failed to add LMB back, drc index %x\n",
470 				       lmb->drc_index);
471 
472 			drmem_remove_lmb_reservation(lmb);
473 		}
474 
475 		rc = -EINVAL;
476 	} else {
477 		for_each_drmem_lmb(lmb) {
478 			if (!drmem_lmb_reserved(lmb))
479 				continue;
480 
481 			dlpar_release_drc(lmb->drc_index);
482 			pr_info("Memory at %llx was hot-removed\n",
483 				lmb->base_addr);
484 
485 			drmem_remove_lmb_reservation(lmb);
486 		}
487 		rc = 0;
488 	}
489 
490 	return rc;
491 }
492 
dlpar_memory_remove_by_index(u32 drc_index)493 static int dlpar_memory_remove_by_index(u32 drc_index)
494 {
495 	struct drmem_lmb *lmb;
496 	int lmb_found;
497 	int rc;
498 
499 	pr_debug("Attempting to hot-remove LMB, drc index %x\n", drc_index);
500 
501 	lmb_found = 0;
502 	for_each_drmem_lmb(lmb) {
503 		if (lmb->drc_index == drc_index) {
504 			lmb_found = 1;
505 			rc = dlpar_remove_lmb(lmb);
506 			if (!rc)
507 				dlpar_release_drc(lmb->drc_index);
508 
509 			break;
510 		}
511 	}
512 
513 	if (!lmb_found) {
514 		pr_debug("Failed to look up LMB for drc index %x\n", drc_index);
515 		rc = -EINVAL;
516 	} else if (rc) {
517 		pr_debug("Failed to hot-remove memory at %llx\n",
518 			 lmb->base_addr);
519 	} else {
520 		pr_debug("Memory at %llx was hot-removed\n", lmb->base_addr);
521 	}
522 
523 	return rc;
524 }
525 
dlpar_memory_readd_by_index(u32 drc_index)526 static int dlpar_memory_readd_by_index(u32 drc_index)
527 {
528 	struct drmem_lmb *lmb;
529 	int lmb_found;
530 	int rc;
531 
532 	pr_info("Attempting to update LMB, drc index %x\n", drc_index);
533 
534 	lmb_found = 0;
535 	for_each_drmem_lmb(lmb) {
536 		if (lmb->drc_index == drc_index) {
537 			lmb_found = 1;
538 			rc = dlpar_remove_lmb(lmb);
539 			if (!rc) {
540 				rc = dlpar_add_lmb(lmb);
541 				if (rc)
542 					dlpar_release_drc(lmb->drc_index);
543 			}
544 			break;
545 		}
546 	}
547 
548 	if (!lmb_found)
549 		rc = -EINVAL;
550 
551 	if (rc)
552 		pr_info("Failed to update memory at %llx\n",
553 			lmb->base_addr);
554 	else
555 		pr_info("Memory at %llx was updated\n", lmb->base_addr);
556 
557 	return rc;
558 }
559 
dlpar_memory_remove_by_ic(u32 lmbs_to_remove,u32 drc_index)560 static int dlpar_memory_remove_by_ic(u32 lmbs_to_remove, u32 drc_index)
561 {
562 	struct drmem_lmb *lmb, *start_lmb, *end_lmb;
563 	int lmbs_available = 0;
564 	int rc;
565 
566 	pr_info("Attempting to hot-remove %u LMB(s) at %x\n",
567 		lmbs_to_remove, drc_index);
568 
569 	if (lmbs_to_remove == 0)
570 		return -EINVAL;
571 
572 	rc = get_lmb_range(drc_index, lmbs_to_remove, &start_lmb, &end_lmb);
573 	if (rc)
574 		return -EINVAL;
575 
576 	/* Validate that there are enough LMBs to satisfy the request */
577 	for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
578 		if (lmb->flags & DRCONF_MEM_RESERVED)
579 			break;
580 
581 		lmbs_available++;
582 	}
583 
584 	if (lmbs_available < lmbs_to_remove)
585 		return -EINVAL;
586 
587 	for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
588 		if (!(lmb->flags & DRCONF_MEM_ASSIGNED))
589 			continue;
590 
591 		rc = dlpar_remove_lmb(lmb);
592 		if (rc)
593 			break;
594 
595 		drmem_mark_lmb_reserved(lmb);
596 	}
597 
598 	if (rc) {
599 		pr_err("Memory indexed-count-remove failed, adding any removed LMBs\n");
600 
601 
602 		for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
603 			if (!drmem_lmb_reserved(lmb))
604 				continue;
605 
606 			rc = dlpar_add_lmb(lmb);
607 			if (rc)
608 				pr_err("Failed to add LMB, drc index %x\n",
609 				       lmb->drc_index);
610 
611 			drmem_remove_lmb_reservation(lmb);
612 		}
613 		rc = -EINVAL;
614 	} else {
615 		for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
616 			if (!drmem_lmb_reserved(lmb))
617 				continue;
618 
619 			dlpar_release_drc(lmb->drc_index);
620 			pr_info("Memory at %llx (drc index %x) was hot-removed\n",
621 				lmb->base_addr, lmb->drc_index);
622 
623 			drmem_remove_lmb_reservation(lmb);
624 		}
625 	}
626 
627 	return rc;
628 }
629 
630 #else
pseries_remove_memblock(unsigned long base,unsigned long memblock_size)631 static inline int pseries_remove_memblock(unsigned long base,
632 					  unsigned long memblock_size)
633 {
634 	return -EOPNOTSUPP;
635 }
pseries_remove_mem_node(struct device_node * np)636 static inline int pseries_remove_mem_node(struct device_node *np)
637 {
638 	return 0;
639 }
dlpar_memory_remove(struct pseries_hp_errorlog * hp_elog)640 static inline int dlpar_memory_remove(struct pseries_hp_errorlog *hp_elog)
641 {
642 	return -EOPNOTSUPP;
643 }
dlpar_remove_lmb(struct drmem_lmb * lmb)644 static int dlpar_remove_lmb(struct drmem_lmb *lmb)
645 {
646 	return -EOPNOTSUPP;
647 }
dlpar_memory_remove_by_count(u32 lmbs_to_remove)648 static int dlpar_memory_remove_by_count(u32 lmbs_to_remove)
649 {
650 	return -EOPNOTSUPP;
651 }
dlpar_memory_remove_by_index(u32 drc_index)652 static int dlpar_memory_remove_by_index(u32 drc_index)
653 {
654 	return -EOPNOTSUPP;
655 }
dlpar_memory_readd_by_index(u32 drc_index)656 static int dlpar_memory_readd_by_index(u32 drc_index)
657 {
658 	return -EOPNOTSUPP;
659 }
660 
dlpar_memory_remove_by_ic(u32 lmbs_to_remove,u32 drc_index)661 static int dlpar_memory_remove_by_ic(u32 lmbs_to_remove, u32 drc_index)
662 {
663 	return -EOPNOTSUPP;
664 }
665 #endif /* CONFIG_MEMORY_HOTREMOVE */
666 
dlpar_add_lmb(struct drmem_lmb * lmb)667 static int dlpar_add_lmb(struct drmem_lmb *lmb)
668 {
669 	unsigned long block_sz;
670 	int nid, rc;
671 
672 	if (lmb->flags & DRCONF_MEM_ASSIGNED)
673 		return -EINVAL;
674 
675 	rc = update_lmb_associativity_index(lmb);
676 	if (rc) {
677 		dlpar_release_drc(lmb->drc_index);
678 		return rc;
679 	}
680 
681 	block_sz = memory_block_size_bytes();
682 
683 	/* Find the node id for this address. */
684 	nid = memory_add_physaddr_to_nid(lmb->base_addr);
685 
686 	/* Add the memory */
687 	rc = __add_memory(nid, lmb->base_addr, block_sz);
688 	if (rc) {
689 		invalidate_lmb_associativity_index(lmb);
690 		return rc;
691 	}
692 
693 	rc = dlpar_online_lmb(lmb);
694 	if (rc) {
695 		__remove_memory(nid, lmb->base_addr, block_sz);
696 		invalidate_lmb_associativity_index(lmb);
697 	} else {
698 		lmb->flags |= DRCONF_MEM_ASSIGNED;
699 	}
700 
701 	return rc;
702 }
703 
dlpar_memory_add_by_count(u32 lmbs_to_add)704 static int dlpar_memory_add_by_count(u32 lmbs_to_add)
705 {
706 	struct drmem_lmb *lmb;
707 	int lmbs_available = 0;
708 	int lmbs_added = 0;
709 	int rc;
710 
711 	pr_info("Attempting to hot-add %d LMB(s)\n", lmbs_to_add);
712 
713 	if (lmbs_to_add == 0)
714 		return -EINVAL;
715 
716 	/* Validate that there are enough LMBs to satisfy the request */
717 	for_each_drmem_lmb(lmb) {
718 		if (!(lmb->flags & DRCONF_MEM_ASSIGNED))
719 			lmbs_available++;
720 
721 		if (lmbs_available == lmbs_to_add)
722 			break;
723 	}
724 
725 	if (lmbs_available < lmbs_to_add)
726 		return -EINVAL;
727 
728 	for_each_drmem_lmb(lmb) {
729 		if (lmb->flags & DRCONF_MEM_ASSIGNED)
730 			continue;
731 
732 		rc = dlpar_acquire_drc(lmb->drc_index);
733 		if (rc)
734 			continue;
735 
736 		rc = dlpar_add_lmb(lmb);
737 		if (rc) {
738 			dlpar_release_drc(lmb->drc_index);
739 			continue;
740 		}
741 
742 		/* Mark this lmb so we can remove it later if all of the
743 		 * requested LMBs cannot be added.
744 		 */
745 		drmem_mark_lmb_reserved(lmb);
746 
747 		lmbs_added++;
748 		if (lmbs_added == lmbs_to_add)
749 			break;
750 	}
751 
752 	if (lmbs_added != lmbs_to_add) {
753 		pr_err("Memory hot-add failed, removing any added LMBs\n");
754 
755 		for_each_drmem_lmb(lmb) {
756 			if (!drmem_lmb_reserved(lmb))
757 				continue;
758 
759 			rc = dlpar_remove_lmb(lmb);
760 			if (rc)
761 				pr_err("Failed to remove LMB, drc index %x\n",
762 				       lmb->drc_index);
763 			else
764 				dlpar_release_drc(lmb->drc_index);
765 
766 			drmem_remove_lmb_reservation(lmb);
767 		}
768 		rc = -EINVAL;
769 	} else {
770 		for_each_drmem_lmb(lmb) {
771 			if (!drmem_lmb_reserved(lmb))
772 				continue;
773 
774 			pr_debug("Memory at %llx (drc index %x) was hot-added\n",
775 				 lmb->base_addr, lmb->drc_index);
776 			drmem_remove_lmb_reservation(lmb);
777 		}
778 		rc = 0;
779 	}
780 
781 	return rc;
782 }
783 
dlpar_memory_add_by_index(u32 drc_index)784 static int dlpar_memory_add_by_index(u32 drc_index)
785 {
786 	struct drmem_lmb *lmb;
787 	int rc, lmb_found;
788 
789 	pr_info("Attempting to hot-add LMB, drc index %x\n", drc_index);
790 
791 	lmb_found = 0;
792 	for_each_drmem_lmb(lmb) {
793 		if (lmb->drc_index == drc_index) {
794 			lmb_found = 1;
795 			rc = dlpar_acquire_drc(lmb->drc_index);
796 			if (!rc) {
797 				rc = dlpar_add_lmb(lmb);
798 				if (rc)
799 					dlpar_release_drc(lmb->drc_index);
800 			}
801 
802 			break;
803 		}
804 	}
805 
806 	if (!lmb_found)
807 		rc = -EINVAL;
808 
809 	if (rc)
810 		pr_info("Failed to hot-add memory, drc index %x\n", drc_index);
811 	else
812 		pr_info("Memory at %llx (drc index %x) was hot-added\n",
813 			lmb->base_addr, drc_index);
814 
815 	return rc;
816 }
817 
dlpar_memory_add_by_ic(u32 lmbs_to_add,u32 drc_index)818 static int dlpar_memory_add_by_ic(u32 lmbs_to_add, u32 drc_index)
819 {
820 	struct drmem_lmb *lmb, *start_lmb, *end_lmb;
821 	int lmbs_available = 0;
822 	int rc;
823 
824 	pr_info("Attempting to hot-add %u LMB(s) at index %x\n",
825 		lmbs_to_add, drc_index);
826 
827 	if (lmbs_to_add == 0)
828 		return -EINVAL;
829 
830 	rc = get_lmb_range(drc_index, lmbs_to_add, &start_lmb, &end_lmb);
831 	if (rc)
832 		return -EINVAL;
833 
834 	/* Validate that the LMBs in this range are not reserved */
835 	for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
836 		if (lmb->flags & DRCONF_MEM_RESERVED)
837 			break;
838 
839 		lmbs_available++;
840 	}
841 
842 	if (lmbs_available < lmbs_to_add)
843 		return -EINVAL;
844 
845 	for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
846 		if (lmb->flags & DRCONF_MEM_ASSIGNED)
847 			continue;
848 
849 		rc = dlpar_acquire_drc(lmb->drc_index);
850 		if (rc)
851 			break;
852 
853 		rc = dlpar_add_lmb(lmb);
854 		if (rc) {
855 			dlpar_release_drc(lmb->drc_index);
856 			break;
857 		}
858 
859 		drmem_mark_lmb_reserved(lmb);
860 	}
861 
862 	if (rc) {
863 		pr_err("Memory indexed-count-add failed, removing any added LMBs\n");
864 
865 		for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
866 			if (!drmem_lmb_reserved(lmb))
867 				continue;
868 
869 			rc = dlpar_remove_lmb(lmb);
870 			if (rc)
871 				pr_err("Failed to remove LMB, drc index %x\n",
872 				       lmb->drc_index);
873 			else
874 				dlpar_release_drc(lmb->drc_index);
875 
876 			drmem_remove_lmb_reservation(lmb);
877 		}
878 		rc = -EINVAL;
879 	} else {
880 		for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
881 			if (!drmem_lmb_reserved(lmb))
882 				continue;
883 
884 			pr_info("Memory at %llx (drc index %x) was hot-added\n",
885 				lmb->base_addr, lmb->drc_index);
886 			drmem_remove_lmb_reservation(lmb);
887 		}
888 	}
889 
890 	return rc;
891 }
892 
dlpar_memory(struct pseries_hp_errorlog * hp_elog)893 int dlpar_memory(struct pseries_hp_errorlog *hp_elog)
894 {
895 	u32 count, drc_index;
896 	int rc;
897 
898 	lock_device_hotplug();
899 
900 	switch (hp_elog->action) {
901 	case PSERIES_HP_ELOG_ACTION_ADD:
902 		switch (hp_elog->id_type) {
903 		case PSERIES_HP_ELOG_ID_DRC_COUNT:
904 			count = hp_elog->_drc_u.drc_count;
905 			rc = dlpar_memory_add_by_count(count);
906 			break;
907 		case PSERIES_HP_ELOG_ID_DRC_INDEX:
908 			drc_index = hp_elog->_drc_u.drc_index;
909 			rc = dlpar_memory_add_by_index(drc_index);
910 			break;
911 		case PSERIES_HP_ELOG_ID_DRC_IC:
912 			count = hp_elog->_drc_u.ic.count;
913 			drc_index = hp_elog->_drc_u.ic.index;
914 			rc = dlpar_memory_add_by_ic(count, drc_index);
915 			break;
916 		default:
917 			rc = -EINVAL;
918 			break;
919 		}
920 
921 		break;
922 	case PSERIES_HP_ELOG_ACTION_REMOVE:
923 		switch (hp_elog->id_type) {
924 		case PSERIES_HP_ELOG_ID_DRC_COUNT:
925 			count = hp_elog->_drc_u.drc_count;
926 			rc = dlpar_memory_remove_by_count(count);
927 			break;
928 		case PSERIES_HP_ELOG_ID_DRC_INDEX:
929 			drc_index = hp_elog->_drc_u.drc_index;
930 			rc = dlpar_memory_remove_by_index(drc_index);
931 			break;
932 		case PSERIES_HP_ELOG_ID_DRC_IC:
933 			count = hp_elog->_drc_u.ic.count;
934 			drc_index = hp_elog->_drc_u.ic.index;
935 			rc = dlpar_memory_remove_by_ic(count, drc_index);
936 			break;
937 		default:
938 			rc = -EINVAL;
939 			break;
940 		}
941 
942 		break;
943 	case PSERIES_HP_ELOG_ACTION_READD:
944 		drc_index = hp_elog->_drc_u.drc_index;
945 		rc = dlpar_memory_readd_by_index(drc_index);
946 		break;
947 	default:
948 		pr_err("Invalid action (%d) specified\n", hp_elog->action);
949 		rc = -EINVAL;
950 		break;
951 	}
952 
953 	if (!rc) {
954 		rtas_hp_event = true;
955 		rc = drmem_update_dt();
956 		rtas_hp_event = false;
957 	}
958 
959 	unlock_device_hotplug();
960 	return rc;
961 }
962 
pseries_add_mem_node(struct device_node * np)963 static int pseries_add_mem_node(struct device_node *np)
964 {
965 	const __be32 *prop;
966 	unsigned long base;
967 	unsigned long lmb_size;
968 	int ret = -EINVAL;
969 	int addr_cells, size_cells;
970 
971 	/*
972 	 * Check to see if we are actually adding memory
973 	 */
974 	if (!of_node_is_type(np, "memory"))
975 		return 0;
976 
977 	/*
978 	 * Find the base and size of the memblock
979 	 */
980 	prop = of_get_property(np, "reg", NULL);
981 	if (!prop)
982 		return ret;
983 
984 	addr_cells = of_n_addr_cells(np);
985 	size_cells = of_n_size_cells(np);
986 	/*
987 	 * "reg" property represents (addr,size) tuple.
988 	 */
989 	base = of_read_number(prop, addr_cells);
990 	prop += addr_cells;
991 	lmb_size = of_read_number(prop, size_cells);
992 
993 	/*
994 	 * Update memory region to represent the memory add
995 	 */
996 	ret = memblock_add(base, lmb_size);
997 	return (ret < 0) ? -EINVAL : 0;
998 }
999 
pseries_update_drconf_memory(struct of_reconfig_data * pr)1000 static int pseries_update_drconf_memory(struct of_reconfig_data *pr)
1001 {
1002 	struct of_drconf_cell_v1 *new_drmem, *old_drmem;
1003 	unsigned long memblock_size;
1004 	u32 entries;
1005 	__be32 *p;
1006 	int i, rc = -EINVAL;
1007 
1008 	if (rtas_hp_event)
1009 		return 0;
1010 
1011 	memblock_size = pseries_memory_block_size();
1012 	if (!memblock_size)
1013 		return -EINVAL;
1014 
1015 	if (!pr->old_prop)
1016 		return 0;
1017 
1018 	p = (__be32 *) pr->old_prop->value;
1019 	if (!p)
1020 		return -EINVAL;
1021 
1022 	/* The first int of the property is the number of lmb's described
1023 	 * by the property. This is followed by an array of of_drconf_cell
1024 	 * entries. Get the number of entries and skip to the array of
1025 	 * of_drconf_cell's.
1026 	 */
1027 	entries = be32_to_cpu(*p++);
1028 	old_drmem = (struct of_drconf_cell_v1 *)p;
1029 
1030 	p = (__be32 *)pr->prop->value;
1031 	p++;
1032 	new_drmem = (struct of_drconf_cell_v1 *)p;
1033 
1034 	for (i = 0; i < entries; i++) {
1035 		if ((be32_to_cpu(old_drmem[i].flags) & DRCONF_MEM_ASSIGNED) &&
1036 		    (!(be32_to_cpu(new_drmem[i].flags) & DRCONF_MEM_ASSIGNED))) {
1037 			rc = pseries_remove_memblock(
1038 				be64_to_cpu(old_drmem[i].base_addr),
1039 						     memblock_size);
1040 			break;
1041 		} else if ((!(be32_to_cpu(old_drmem[i].flags) &
1042 			    DRCONF_MEM_ASSIGNED)) &&
1043 			    (be32_to_cpu(new_drmem[i].flags) &
1044 			    DRCONF_MEM_ASSIGNED)) {
1045 			rc = memblock_add(be64_to_cpu(old_drmem[i].base_addr),
1046 					  memblock_size);
1047 			rc = (rc < 0) ? -EINVAL : 0;
1048 			break;
1049 		}
1050 	}
1051 	return rc;
1052 }
1053 
pseries_memory_notifier(struct notifier_block * nb,unsigned long action,void * data)1054 static int pseries_memory_notifier(struct notifier_block *nb,
1055 				   unsigned long action, void *data)
1056 {
1057 	struct of_reconfig_data *rd = data;
1058 	int err = 0;
1059 
1060 	switch (action) {
1061 	case OF_RECONFIG_ATTACH_NODE:
1062 		err = pseries_add_mem_node(rd->dn);
1063 		break;
1064 	case OF_RECONFIG_DETACH_NODE:
1065 		err = pseries_remove_mem_node(rd->dn);
1066 		break;
1067 	case OF_RECONFIG_UPDATE_PROPERTY:
1068 		if (!strcmp(rd->prop->name, "ibm,dynamic-memory"))
1069 			err = pseries_update_drconf_memory(rd);
1070 		break;
1071 	}
1072 	return notifier_from_errno(err);
1073 }
1074 
1075 static struct notifier_block pseries_mem_nb = {
1076 	.notifier_call = pseries_memory_notifier,
1077 };
1078 
pseries_memory_hotplug_init(void)1079 static int __init pseries_memory_hotplug_init(void)
1080 {
1081 	if (firmware_has_feature(FW_FEATURE_LPAR))
1082 		of_reconfig_notifier_register(&pseries_mem_nb);
1083 
1084 	return 0;
1085 }
1086 machine_device_initcall(pseries, pseries_memory_hotplug_init);
1087