[PATCH v7 2/8] x86/crash: Introduce new options to support cpu and memory hotplug
Sourabh Jain
sourabhjain at linux.ibm.com
Thu Apr 28 23:41:56 PDT 2022
On 26/04/22 20:09, Eric DeVolder wrote:
>
>
> On 4/25/22 23:21, Sourabh Jain wrote:
>>
>> On 13/04/22 22:12, Eric DeVolder wrote:
>>> CRASH_HOTPLUG is to enable cpu and memory hotplug support of crash.
>>>
>>> CRASH_HOTPLUG_ELFCOREHDR_SZ is used to specify the maximum size of
>>> the elfcorehdr buffer/segment.
>>>
>>> This is a preparation for later usage.
>>>
>>> Signed-off-by: Eric DeVolder <eric.devolder at oracle.com>
>>> Acked-by: Baoquan He <bhe at redhat.com>
>>> ---
>>> arch/x86/Kconfig | 26 ++++++++++++++++++++++++++
>>> 1 file changed, 26 insertions(+)
>>>
>>> diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig
>>> index b0142e01002e..f7b92ee1bcc7 100644
>>> --- a/arch/x86/Kconfig
>>> +++ b/arch/x86/Kconfig
>>> @@ -2072,6 +2072,32 @@ config CRASH_DUMP
>>> (CONFIG_RELOCATABLE=y).
>>> For more details see Documentation/admin-guide/kdump/kdump.rst
>>> +config CRASH_HOTPLUG
>>> + bool "kernel updates of crash elfcorehdr"
>>> + depends on CRASH_DUMP && (HOTPLUG_CPU || MEMORY_HOTPLUG) &&
>>> KEXEC_FILE
>>> + help
>>> + Enable the kernel to update the crash elfcorehdr (which contains
>>> + the list of CPUs and memory regions) directly when hot
>>> plug/unplug
>>> + of CPUs or memory. Otherwise userspace must monitor these hot
>>> + plug/unplug change notifications via udev in order to
>>> + unload-then-reload the crash kernel so that the list of CPUs and
>>> + memory regions is kept up-to-date. Note that the udev CPU and
>>> + memory change notifications still occur (however, userspace
>>> is not
>>> + required to monitor for crash dump purposes).
>>> +
>>> +config CRASH_HOTPLUG_ELFCOREHDR_SZ
>>> + depends on CRASH_HOTPLUG
>>> + int
>>> + default 131072
>>> + help
>>> + Specify the maximum size of the elfcorehdr buffer/segment.
>>> + The 128KiB default is sized so that it can accommodate 2048
>>> + Elf64_Phdr, where each Phdr represents either a CPU or a
>>> + region of memory.
>>> + For example, this size can accommodate a machine with up to 1024
>>> + CPUs and up to 1024 memory regions, eg. as represented by the
>>> + 'System RAM' entries in /proc/iomem.
>>
>> Is it possible to get rid of CRASH_HOTPLUG_ELFCOREHDR_SZ?
> At the moment, I do not think so. The idea behind this value is to
> represent the largest number of CPUs and memory regions possible in
> the system. Today there is NR_CPUS which could be used for CPUs, but
> there isn't a similar value for memory. I also am not aware of a
> kernel variable that could be utilized to represent the maximum number
> of memory regions. If there is, please let me know!
>>
>> How about finding the additional buffer space needed for future CPU
>> and memory
>> add during the kdump load? Not sure about the feasibility of doing
>> this in
>> kexec tool (userspace).
>
> I may not understand what you are asking, but the x86 code, for
> kexec_file_load, does in fact allocate all the space needed (currently
> via CRASH_HOTPLUG_ELFCOREHDR_SZ) upon kdump load.
>
> For kexec_load, I've had no problem asking the kexec tool to allocate
> a larger piece of memory for the elfcorehdr. But it is the same
> problem as CRASH_HOTPLUG_ELFCOREHDR_SZ; how big? In my workspace I
> tell kexec tool how big. If there are sysfs visible values for NR_CPU
> and memory, then we could have kexec pull those and compute.
Yeah dynamic calculation for PT_LOAD sections needed for possible memory
may not be straightforward. But still I did not get the rational for
limiting the possible PT_LOAD sections or memory ranges to only 1024.
Although in kexec tool the max memory ranges for x86 is 32K.
commit 1bc7bc7649fa29d95c98f6a6d8dd2f08734a865c
Author: David Hildenbrand <david at redhat.com>
Date: Tue Mar 23 11:01:10 2021 +0100
crashdump/x86: increase CRASH_MAX_MEMORY_RANGES to 32k
virtio-mem in Linux adds/removes individual memory blocks (e.g., 128 MB
each). Linux merges adjacent memory blocks added by virtio-mem
devices, but
we can still end up with a very sparse memory layout when unplugging
memory in corner cases.
Let's increase the maximum number of crash memory ranges from ~2k
to 32k.
32k should be sufficient for a very long time.
e_phnum field in the header is 16 bits wide, so we can fit a maximum of
~64k entries in there, shared with other entries (i.e., CPU).
Therefore,
using up to 32k memory ranges is fine. (if we ever need more than ~64k,
Do you see any issue if we increase the memory range count to 32K?
Thanks,
Sourabh Jain
More information about the kexec
mailing list