1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * pseries Memory Hotplug infrastructure.
5 * Copyright (C) 2008 Badari Pulavarty, IBM Corporation
8 #define pr_fmt(fmt) "pseries-hotplug-mem: " fmt
11 #include <linux/of_address.h>
12 #include <linux/memblock.h>
13 #include <linux/memory.h>
14 #include <linux/memory_hotplug.h>
15 #include <linux/slab.h>
17 #include <asm/firmware.h>
18 #include <asm/machdep.h>
20 #include <asm/sparsemem.h>
21 #include <asm/fadump.h>
22 #include <asm/drmem.h>
25 unsigned long pseries_memory_block_size(void)
27 struct device_node *np;
28 u64 memblock_size = MIN_MEMORY_BLOCK_SIZE;
31 np = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory");
37 size_cells = of_n_size_cells(np);
39 prop = of_get_property(np, "ibm,lmb-size", &len);
40 if (prop && len >= size_cells * sizeof(__be32))
41 memblock_size = of_read_number(prop, size_cells);
44 } else if (machine_is(pseries)) {
45 /* This fallback really only applies to pseries */
46 unsigned int memzero_size = 0;
48 np = of_find_node_by_path("/memory@0");
50 if (!of_address_to_resource(np, 0, &r))
51 memzero_size = resource_size(&r);
56 /* We now know the size of memory@0, use this to find
57 * the first memoryblock and get its size.
61 sprintf(buf, "/memory@%x", memzero_size);
62 np = of_find_node_by_path(buf);
64 if (!of_address_to_resource(np, 0, &r))
65 memblock_size = resource_size(&r);
73 static void dlpar_free_property(struct property *prop)
80 static struct property *dlpar_clone_property(struct property *prop,
83 struct property *new_prop;
85 new_prop = kzalloc(sizeof(*new_prop), GFP_KERNEL);
89 new_prop->name = kstrdup(prop->name, GFP_KERNEL);
90 new_prop->value = kzalloc(prop_size, GFP_KERNEL);
91 if (!new_prop->name || !new_prop->value) {
92 dlpar_free_property(new_prop);
96 memcpy(new_prop->value, prop->value, prop->length);
97 new_prop->length = prop_size;
99 of_property_set_flag(new_prop, OF_DYNAMIC);
103 static bool find_aa_index(struct device_node *dr_node,
104 struct property *ala_prop,
105 const u32 *lmb_assoc, u32 *aa_index)
107 u32 *assoc_arrays, new_prop_size;
108 struct property *new_prop;
109 int aa_arrays, aa_array_entries, aa_array_sz;
113 * The ibm,associativity-lookup-arrays property is defined to be
114 * a 32-bit value specifying the number of associativity arrays
115 * followed by a 32-bitvalue specifying the number of entries per
116 * array, followed by the associativity arrays.
118 assoc_arrays = ala_prop->value;
120 aa_arrays = be32_to_cpu(assoc_arrays[0]);
121 aa_array_entries = be32_to_cpu(assoc_arrays[1]);
122 aa_array_sz = aa_array_entries * sizeof(u32);
124 for (i = 0; i < aa_arrays; i++) {
125 index = (i * aa_array_entries) + 2;
127 if (memcmp(&assoc_arrays[index], &lmb_assoc[1], aa_array_sz))
134 new_prop_size = ala_prop->length + aa_array_sz;
135 new_prop = dlpar_clone_property(ala_prop, new_prop_size);
139 assoc_arrays = new_prop->value;
141 /* increment the number of entries in the lookup array */
142 assoc_arrays[0] = cpu_to_be32(aa_arrays + 1);
144 /* copy the new associativity into the lookup array */
145 index = aa_arrays * aa_array_entries + 2;
146 memcpy(&assoc_arrays[index], &lmb_assoc[1], aa_array_sz);
148 of_update_property(dr_node, new_prop);
151 * The associativity lookup array index for this lmb is
152 * number of entries - 1 since we added its associativity
153 * to the end of the lookup array.
155 *aa_index = be32_to_cpu(assoc_arrays[0]) - 1;
159 static int update_lmb_associativity_index(struct drmem_lmb *lmb)
161 struct device_node *parent, *lmb_node, *dr_node;
162 struct property *ala_prop;
163 const u32 *lmb_assoc;
167 parent = of_find_node_by_path("/");
171 lmb_node = dlpar_configure_connector(cpu_to_be32(lmb->drc_index),
177 lmb_assoc = of_get_property(lmb_node, "ibm,associativity", NULL);
179 dlpar_free_cc_nodes(lmb_node);
183 update_numa_distance(lmb_node);
185 dr_node = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory");
187 dlpar_free_cc_nodes(lmb_node);
191 ala_prop = of_find_property(dr_node, "ibm,associativity-lookup-arrays",
194 of_node_put(dr_node);
195 dlpar_free_cc_nodes(lmb_node);
199 found = find_aa_index(dr_node, ala_prop, lmb_assoc, &aa_index);
201 of_node_put(dr_node);
202 dlpar_free_cc_nodes(lmb_node);
205 pr_err("Could not find LMB associativity\n");
209 lmb->aa_index = aa_index;
213 static struct memory_block *lmb_to_memblock(struct drmem_lmb *lmb)
215 unsigned long section_nr;
216 struct mem_section *mem_sect;
217 struct memory_block *mem_block;
219 section_nr = pfn_to_section_nr(PFN_DOWN(lmb->base_addr));
220 mem_sect = __nr_to_section(section_nr);
222 mem_block = find_memory_block(mem_sect);
226 static int get_lmb_range(u32 drc_index, int n_lmbs,
227 struct drmem_lmb **start_lmb,
228 struct drmem_lmb **end_lmb)
230 struct drmem_lmb *lmb, *start, *end;
231 struct drmem_lmb *limit;
234 for_each_drmem_lmb(lmb) {
235 if (lmb->drc_index == drc_index) {
244 end = &start[n_lmbs];
246 limit = &drmem_info->lmbs[drmem_info->n_lmbs];
255 static int dlpar_change_lmb_state(struct drmem_lmb *lmb, bool online)
257 struct memory_block *mem_block;
260 mem_block = lmb_to_memblock(lmb);
264 if (online && mem_block->dev.offline)
265 rc = device_online(&mem_block->dev);
266 else if (!online && !mem_block->dev.offline)
267 rc = device_offline(&mem_block->dev);
271 put_device(&mem_block->dev);
276 static int dlpar_online_lmb(struct drmem_lmb *lmb)
278 return dlpar_change_lmb_state(lmb, true);
281 #ifdef CONFIG_MEMORY_HOTREMOVE
282 static int dlpar_offline_lmb(struct drmem_lmb *lmb)
284 return dlpar_change_lmb_state(lmb, false);
287 static int pseries_remove_memblock(unsigned long base, unsigned long memblock_size)
289 unsigned long block_sz, start_pfn;
290 int sections_per_block;
293 start_pfn = base >> PAGE_SHIFT;
295 lock_device_hotplug();
297 if (!pfn_valid(start_pfn))
300 block_sz = pseries_memory_block_size();
301 sections_per_block = block_sz / MIN_MEMORY_BLOCK_SIZE;
302 nid = memory_add_physaddr_to_nid(base);
304 for (i = 0; i < sections_per_block; i++) {
305 __remove_memory(nid, base, MIN_MEMORY_BLOCK_SIZE);
306 base += MIN_MEMORY_BLOCK_SIZE;
310 /* Update memory regions for memory remove */
311 memblock_remove(base, memblock_size);
312 unlock_device_hotplug();
316 static int pseries_remove_mem_node(struct device_node *np)
320 unsigned long lmb_size;
322 int addr_cells, size_cells;
325 * Check to see if we are actually removing memory
327 if (!of_node_is_type(np, "memory"))
331 * Find the base address and size of the memblock
333 prop = of_get_property(np, "reg", NULL);
337 addr_cells = of_n_addr_cells(np);
338 size_cells = of_n_size_cells(np);
341 * "reg" property represents (addr,size) tuple.
343 base = of_read_number(prop, addr_cells);
345 lmb_size = of_read_number(prop, size_cells);
347 pseries_remove_memblock(base, lmb_size);
351 static bool lmb_is_removable(struct drmem_lmb *lmb)
353 if ((lmb->flags & DRCONF_MEM_RESERVED) ||
354 !(lmb->flags & DRCONF_MEM_ASSIGNED))
357 #ifdef CONFIG_FA_DUMP
359 * Don't hot-remove memory that falls in fadump boot memory area
360 * and memory that is reserved for capturing old kernel memory.
362 if (is_fadump_memory_area(lmb->base_addr, memory_block_size_bytes()))
365 /* device_offline() will determine if we can actually remove this lmb */
369 static int dlpar_add_lmb(struct drmem_lmb *);
371 static int dlpar_remove_lmb(struct drmem_lmb *lmb)
373 struct memory_block *mem_block;
374 unsigned long block_sz;
377 if (!lmb_is_removable(lmb))
380 mem_block = lmb_to_memblock(lmb);
381 if (mem_block == NULL)
384 rc = dlpar_offline_lmb(lmb);
386 put_device(&mem_block->dev);
390 block_sz = pseries_memory_block_size();
392 __remove_memory(mem_block->nid, lmb->base_addr, block_sz);
393 put_device(&mem_block->dev);
395 /* Update memory regions for memory remove */
396 memblock_remove(lmb->base_addr, block_sz);
398 invalidate_lmb_associativity_index(lmb);
399 lmb->flags &= ~DRCONF_MEM_ASSIGNED;
404 static int dlpar_memory_remove_by_count(u32 lmbs_to_remove)
406 struct drmem_lmb *lmb;
407 int lmbs_reserved = 0;
408 int lmbs_available = 0;
411 pr_info("Attempting to hot-remove %d LMB(s)\n", lmbs_to_remove);
413 if (lmbs_to_remove == 0)
416 /* Validate that there are enough LMBs to satisfy the request */
417 for_each_drmem_lmb(lmb) {
418 if (lmb_is_removable(lmb))
421 if (lmbs_available == lmbs_to_remove)
425 if (lmbs_available < lmbs_to_remove) {
426 pr_info("Not enough LMBs available (%d of %d) to satisfy request\n",
427 lmbs_available, lmbs_to_remove);
431 for_each_drmem_lmb(lmb) {
432 rc = dlpar_remove_lmb(lmb);
436 /* Mark this lmb so we can add it later if all of the
437 * requested LMBs cannot be removed.
439 drmem_mark_lmb_reserved(lmb);
442 if (lmbs_reserved == lmbs_to_remove)
446 if (lmbs_reserved != lmbs_to_remove) {
447 pr_err("Memory hot-remove failed, adding LMB's back\n");
449 for_each_drmem_lmb(lmb) {
450 if (!drmem_lmb_reserved(lmb))
453 rc = dlpar_add_lmb(lmb);
455 pr_err("Failed to add LMB back, drc index %x\n",
458 drmem_remove_lmb_reservation(lmb);
461 if (lmbs_reserved == 0)
467 for_each_drmem_lmb(lmb) {
468 if (!drmem_lmb_reserved(lmb))
471 dlpar_release_drc(lmb->drc_index);
472 pr_info("Memory at %llx was hot-removed\n",
475 drmem_remove_lmb_reservation(lmb);
478 if (lmbs_reserved == 0)
487 static int dlpar_memory_remove_by_index(u32 drc_index)
489 struct drmem_lmb *lmb;
493 pr_debug("Attempting to hot-remove LMB, drc index %x\n", drc_index);
496 for_each_drmem_lmb(lmb) {
497 if (lmb->drc_index == drc_index) {
499 rc = dlpar_remove_lmb(lmb);
501 dlpar_release_drc(lmb->drc_index);
511 pr_debug("Failed to hot-remove memory at %llx\n",
514 pr_debug("Memory at %llx was hot-removed\n", lmb->base_addr);
519 static int dlpar_memory_remove_by_ic(u32 lmbs_to_remove, u32 drc_index)
521 struct drmem_lmb *lmb, *start_lmb, *end_lmb;
524 pr_info("Attempting to hot-remove %u LMB(s) at %x\n",
525 lmbs_to_remove, drc_index);
527 if (lmbs_to_remove == 0)
530 rc = get_lmb_range(drc_index, lmbs_to_remove, &start_lmb, &end_lmb);
535 * Validate that all LMBs in range are not reserved. Note that it
536 * is ok if they are !ASSIGNED since our goal here is to remove the
537 * LMB range, regardless of whether some LMBs were already removed
538 * by any other reason.
540 * This is a contrast to what is done in remove_by_count() where we
541 * check for both RESERVED and !ASSIGNED (via lmb_is_removable()),
542 * because we want to remove a fixed amount of LMBs in that function.
544 for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
545 if (lmb->flags & DRCONF_MEM_RESERVED) {
546 pr_err("Memory at %llx (drc index %x) is reserved\n",
547 lmb->base_addr, lmb->drc_index);
552 for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
554 * dlpar_remove_lmb() will error out if the LMB is already
555 * !ASSIGNED, but this case is a no-op for us.
557 if (!(lmb->flags & DRCONF_MEM_ASSIGNED))
560 rc = dlpar_remove_lmb(lmb);
564 drmem_mark_lmb_reserved(lmb);
568 pr_err("Memory indexed-count-remove failed, adding any removed LMBs\n");
571 for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
572 if (!drmem_lmb_reserved(lmb))
576 * Setting the isolation state of an UNISOLATED/CONFIGURED
577 * device to UNISOLATE is a no-op, but the hypervisor can
578 * use it as a hint that the LMB removal failed.
580 dlpar_unisolate_drc(lmb->drc_index);
582 rc = dlpar_add_lmb(lmb);
584 pr_err("Failed to add LMB, drc index %x\n",
587 drmem_remove_lmb_reservation(lmb);
591 for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
592 if (!drmem_lmb_reserved(lmb))
595 dlpar_release_drc(lmb->drc_index);
596 pr_info("Memory at %llx (drc index %x) was hot-removed\n",
597 lmb->base_addr, lmb->drc_index);
599 drmem_remove_lmb_reservation(lmb);
607 static inline int pseries_remove_memblock(unsigned long base,
608 unsigned long memblock_size)
612 static inline int pseries_remove_mem_node(struct device_node *np)
616 static int dlpar_remove_lmb(struct drmem_lmb *lmb)
620 static int dlpar_memory_remove_by_count(u32 lmbs_to_remove)
624 static int dlpar_memory_remove_by_index(u32 drc_index)
629 static int dlpar_memory_remove_by_ic(u32 lmbs_to_remove, u32 drc_index)
633 #endif /* CONFIG_MEMORY_HOTREMOVE */
635 static int dlpar_add_lmb(struct drmem_lmb *lmb)
637 unsigned long block_sz;
640 if (lmb->flags & DRCONF_MEM_ASSIGNED)
643 rc = update_lmb_associativity_index(lmb);
645 dlpar_release_drc(lmb->drc_index);
649 block_sz = memory_block_size_bytes();
651 /* Find the node id for this LMB. Fake one if necessary. */
652 nid = of_drconf_to_nid_single(lmb);
653 if (nid < 0 || !node_possible(nid))
654 nid = first_online_node;
657 rc = __add_memory(nid, lmb->base_addr, block_sz, MHP_NONE);
659 invalidate_lmb_associativity_index(lmb);
663 rc = dlpar_online_lmb(lmb);
665 __remove_memory(nid, lmb->base_addr, block_sz);
666 invalidate_lmb_associativity_index(lmb);
668 lmb->flags |= DRCONF_MEM_ASSIGNED;
674 static int dlpar_memory_add_by_count(u32 lmbs_to_add)
676 struct drmem_lmb *lmb;
677 int lmbs_available = 0;
678 int lmbs_reserved = 0;
681 pr_info("Attempting to hot-add %d LMB(s)\n", lmbs_to_add);
683 if (lmbs_to_add == 0)
686 /* Validate that there are enough LMBs to satisfy the request */
687 for_each_drmem_lmb(lmb) {
688 if (lmb->flags & DRCONF_MEM_RESERVED)
691 if (!(lmb->flags & DRCONF_MEM_ASSIGNED))
694 if (lmbs_available == lmbs_to_add)
698 if (lmbs_available < lmbs_to_add)
701 for_each_drmem_lmb(lmb) {
702 if (lmb->flags & DRCONF_MEM_ASSIGNED)
705 rc = dlpar_acquire_drc(lmb->drc_index);
709 rc = dlpar_add_lmb(lmb);
711 dlpar_release_drc(lmb->drc_index);
715 /* Mark this lmb so we can remove it later if all of the
716 * requested LMBs cannot be added.
718 drmem_mark_lmb_reserved(lmb);
720 if (lmbs_reserved == lmbs_to_add)
724 if (lmbs_reserved != lmbs_to_add) {
725 pr_err("Memory hot-add failed, removing any added LMBs\n");
727 for_each_drmem_lmb(lmb) {
728 if (!drmem_lmb_reserved(lmb))
731 rc = dlpar_remove_lmb(lmb);
733 pr_err("Failed to remove LMB, drc index %x\n",
736 dlpar_release_drc(lmb->drc_index);
738 drmem_remove_lmb_reservation(lmb);
741 if (lmbs_reserved == 0)
746 for_each_drmem_lmb(lmb) {
747 if (!drmem_lmb_reserved(lmb))
750 pr_debug("Memory at %llx (drc index %x) was hot-added\n",
751 lmb->base_addr, lmb->drc_index);
752 drmem_remove_lmb_reservation(lmb);
755 if (lmbs_reserved == 0)
764 static int dlpar_memory_add_by_index(u32 drc_index)
766 struct drmem_lmb *lmb;
769 pr_info("Attempting to hot-add LMB, drc index %x\n", drc_index);
772 for_each_drmem_lmb(lmb) {
773 if (lmb->drc_index == drc_index) {
775 rc = dlpar_acquire_drc(lmb->drc_index);
777 rc = dlpar_add_lmb(lmb);
779 dlpar_release_drc(lmb->drc_index);
790 pr_info("Failed to hot-add memory, drc index %x\n", drc_index);
792 pr_info("Memory at %llx (drc index %x) was hot-added\n",
793 lmb->base_addr, drc_index);
798 static int dlpar_memory_add_by_ic(u32 lmbs_to_add, u32 drc_index)
800 struct drmem_lmb *lmb, *start_lmb, *end_lmb;
803 pr_info("Attempting to hot-add %u LMB(s) at index %x\n",
804 lmbs_to_add, drc_index);
806 if (lmbs_to_add == 0)
809 rc = get_lmb_range(drc_index, lmbs_to_add, &start_lmb, &end_lmb);
813 /* Validate that the LMBs in this range are not reserved */
814 for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
815 /* Fail immediately if the whole range can't be hot-added */
816 if (lmb->flags & DRCONF_MEM_RESERVED) {
817 pr_err("Memory at %llx (drc index %x) is reserved\n",
818 lmb->base_addr, lmb->drc_index);
823 for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
824 if (lmb->flags & DRCONF_MEM_ASSIGNED)
827 rc = dlpar_acquire_drc(lmb->drc_index);
831 rc = dlpar_add_lmb(lmb);
833 dlpar_release_drc(lmb->drc_index);
837 drmem_mark_lmb_reserved(lmb);
841 pr_err("Memory indexed-count-add failed, removing any added LMBs\n");
843 for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
844 if (!drmem_lmb_reserved(lmb))
847 rc = dlpar_remove_lmb(lmb);
849 pr_err("Failed to remove LMB, drc index %x\n",
852 dlpar_release_drc(lmb->drc_index);
854 drmem_remove_lmb_reservation(lmb);
858 for_each_drmem_lmb_in_range(lmb, start_lmb, end_lmb) {
859 if (!drmem_lmb_reserved(lmb))
862 pr_info("Memory at %llx (drc index %x) was hot-added\n",
863 lmb->base_addr, lmb->drc_index);
864 drmem_remove_lmb_reservation(lmb);
871 int dlpar_memory(struct pseries_hp_errorlog *hp_elog)
873 u32 count, drc_index;
876 lock_device_hotplug();
878 switch (hp_elog->action) {
879 case PSERIES_HP_ELOG_ACTION_ADD:
880 switch (hp_elog->id_type) {
881 case PSERIES_HP_ELOG_ID_DRC_COUNT:
882 count = hp_elog->_drc_u.drc_count;
883 rc = dlpar_memory_add_by_count(count);
885 case PSERIES_HP_ELOG_ID_DRC_INDEX:
886 drc_index = hp_elog->_drc_u.drc_index;
887 rc = dlpar_memory_add_by_index(drc_index);
889 case PSERIES_HP_ELOG_ID_DRC_IC:
890 count = hp_elog->_drc_u.ic.count;
891 drc_index = hp_elog->_drc_u.ic.index;
892 rc = dlpar_memory_add_by_ic(count, drc_index);
900 case PSERIES_HP_ELOG_ACTION_REMOVE:
901 switch (hp_elog->id_type) {
902 case PSERIES_HP_ELOG_ID_DRC_COUNT:
903 count = hp_elog->_drc_u.drc_count;
904 rc = dlpar_memory_remove_by_count(count);
906 case PSERIES_HP_ELOG_ID_DRC_INDEX:
907 drc_index = hp_elog->_drc_u.drc_index;
908 rc = dlpar_memory_remove_by_index(drc_index);
910 case PSERIES_HP_ELOG_ID_DRC_IC:
911 count = hp_elog->_drc_u.ic.count;
912 drc_index = hp_elog->_drc_u.ic.index;
913 rc = dlpar_memory_remove_by_ic(count, drc_index);
922 pr_err("Invalid action (%d) specified\n", hp_elog->action);
928 rc = drmem_update_dt();
930 unlock_device_hotplug();
934 static int pseries_add_mem_node(struct device_node *np)
938 unsigned long lmb_size;
940 int addr_cells, size_cells;
943 * Check to see if we are actually adding memory
945 if (!of_node_is_type(np, "memory"))
949 * Find the base and size of the memblock
951 prop = of_get_property(np, "reg", NULL);
955 addr_cells = of_n_addr_cells(np);
956 size_cells = of_n_size_cells(np);
958 * "reg" property represents (addr,size) tuple.
960 base = of_read_number(prop, addr_cells);
962 lmb_size = of_read_number(prop, size_cells);
965 * Update memory region to represent the memory add
967 ret = memblock_add(base, lmb_size);
968 return (ret < 0) ? -EINVAL : 0;
971 static int pseries_memory_notifier(struct notifier_block *nb,
972 unsigned long action, void *data)
974 struct of_reconfig_data *rd = data;
978 case OF_RECONFIG_ATTACH_NODE:
979 err = pseries_add_mem_node(rd->dn);
981 case OF_RECONFIG_DETACH_NODE:
982 err = pseries_remove_mem_node(rd->dn);
984 case OF_RECONFIG_UPDATE_PROPERTY:
985 if (!strcmp(rd->dn->name,
986 "ibm,dynamic-reconfiguration-memory"))
987 drmem_update_lmbs(rd->prop);
989 return notifier_from_errno(err);
992 static struct notifier_block pseries_mem_nb = {
993 .notifier_call = pseries_memory_notifier,
996 static int __init pseries_memory_hotplug_init(void)
998 if (firmware_has_feature(FW_FEATURE_LPAR))
999 of_reconfig_notifier_register(&pseries_mem_nb);
1003 machine_device_initcall(pseries, pseries_memory_hotplug_init);