[PATCH] vmcore: call remap_pfn_range() separately for respective partial pages
HATAYAMA Daisuke
d.hatayama at jp.fujitsu.com
Thu Nov 28 03:48:02 EST 2013
Hello Vivek,
Here is a patch set for mmap failure for /proc/vmcore.
Could you try to use this on the problematic system?
This patch doesn't copy partial pages to the 2nd kernel, only prepares
vmcore objects for respective partial pages to invoke remap_pfn_range()
for individual partial pages.
>From c83dddd23be2a2972dcb3f252598c39abfa23078 Mon Sep 17 00:00:00 2001
From: HATAYAMA Daisuke <d.hatayama at jp.fujitsu.com>
Date: Thu, 28 Nov 2013 14:51:22 +0900
Subject: [PATCH] vmcore: call remap_pfn_range() separately for respective
partial pages
Acording to the report by Vivek in
https://lkml.org/lkml/2013/11/13/439, on some specific systems, some
of the System RAM ranges don't end at page boundary and the later part
of the same page is used for some kind of ACPI data. As a result,
remap_pfn_range() to the partial page failed if mapping range covers a
boundary of the System RAM part and the ACPI data part in the partial
page, due to the detection of different cache types in
track_pfn_remap().
To resolve the issue, call remap_pfn_range() separately for respective
partial pages, not for multiple consequtive pages that don't either
start or end at page boundary, by creating vmcore objects for
respective partial pages.
This patch never changes shape of /proc/vmcore visible from user-land.
Reported-by: Vivek Goyal <vgoyal at redhat.com>
Signed-off-by: HATAYAMA Daisuke <d.hatayama at jp.fujitsu.com>
---
fs/proc/vmcore.c | 108 ++++++++++++++++++++++++++++++++++++++++++-------------
1 file changed, 84 insertions(+), 24 deletions(-)
diff --git a/fs/proc/vmcore.c b/fs/proc/vmcore.c
index 9100d69..e396a1d 100644
--- a/fs/proc/vmcore.c
+++ b/fs/proc/vmcore.c
@@ -816,26 +816,56 @@ static int __init process_ptload_program_headers_elf64(char *elfptr,
vmcore_off = elfsz + elfnotes_sz;
for (i = 0; i < ehdr_ptr->e_phnum; i++, phdr_ptr++) {
- u64 paddr, start, end, size;
+ u64 start, end, size, rest;
+ u64 start_up, start_down, end_up, end_down;
if (phdr_ptr->p_type != PT_LOAD)
continue;
- paddr = phdr_ptr->p_offset;
- start = rounddown(paddr, PAGE_SIZE);
- end = roundup(paddr + phdr_ptr->p_memsz, PAGE_SIZE);
- size = end - start;
+ start = phdr_ptr->p_offset;
+ start_up = roundup(start, PAGE_SIZE);
+ start_down = rounddown(start, PAGE_SIZE);
+ end = phdr_ptr->p_offset + phdr_ptr->p_memsz;
+ end_up = roundup(end, PAGE_SIZE);
+ end_down = rounddown(end, PAGE_SIZE);
+ size = end_up - start_down;
+ rest = phdr_ptr->p_memsz;
+
+ if (!PAGE_ALIGNED(start)) {
+ new = get_new_element();
+ if (!new)
+ return -ENOMEM;
+ new->paddr = start_down;
+ new->size = PAGE_SIZE;
+ list_add_tail(&new->list, vc_list);
+ rest -= min(start_up, end) - start;
+ }
/* Add this contiguous chunk of memory to vmcore list.*/
- new = get_new_element();
- if (!new)
- return -ENOMEM;
- new->paddr = start;
- new->size = size;
- list_add_tail(&new->list, vc_list);
+ if (rest > 0 && start_up < end_down) {
+ new = get_new_element();
+ if (!new)
+ return -ENOMEM;
+ new->paddr = start_up;
+ new->size = end_down - start_up;
+ list_add_tail(&new->list, vc_list);
+ rest -= end_down - start_up;
+ }
+
+ if (rest > 0) {
+ new = get_new_element();
+ if (!new)
+ return -ENOMEM;
+ new->paddr = end_down;
+ new->size = PAGE_SIZE;
+ list_add_tail(&new->list, vc_list);
+ rest -= end - end_down;
+ }
+
+ WARN_ON(rest > 0);
/* Update the program header offset. */
- phdr_ptr->p_offset = vmcore_off + (paddr - start);
+ phdr_ptr->p_offset = vmcore_off + (start - start_down);
vmcore_off = vmcore_off + size;
}
return 0;
@@ -859,26 +889,56 @@ static int __init process_ptload_program_headers_elf32(char *elfptr,
vmcore_off = elfsz + elfnotes_sz;
for (i = 0; i < ehdr_ptr->e_phnum; i++, phdr_ptr++) {
- u64 paddr, start, end, size;
+ u64 start, end, size, rest;
+ u64 start_up, start_down, end_up, end_down;
if (phdr_ptr->p_type != PT_LOAD)
continue;
- paddr = phdr_ptr->p_offset;
- start = rounddown(paddr, PAGE_SIZE);
- end = roundup(paddr + phdr_ptr->p_memsz, PAGE_SIZE);
- size = end - start;
+ start = phdr_ptr->p_offset;
+ start_up = roundup(start, PAGE_SIZE);
+ start_down = rounddown(start, PAGE_SIZE);
+ end = phdr_ptr->p_offset + phdr_ptr->p_memsz;
+ end_up = roundup(end, PAGE_SIZE);
+ end_down = rounddown(end, PAGE_SIZE);
+ rest = phdr_ptr->p_memsz;
+ size = end_up - start_down;
+
+ if (!PAGE_ALIGNED(start)) {
+ new = get_new_element();
+ if (!new)
+ return -ENOMEM;
+ new->paddr = start_down;
+ new->size = PAGE_SIZE;
+ list_add_tail(&new->list, vc_list);
+ rest -= min(start_up, end) - start;
+ }
/* Add this contiguous chunk of memory to vmcore list.*/
- new = get_new_element();
- if (!new)
- return -ENOMEM;
- new->paddr = start;
- new->size = size;
- list_add_tail(&new->list, vc_list);
+ if (rest > 0 && start_up < end_down) {
+ new = get_new_element();
+ if (!new)
+ return -ENOMEM;
+ new->paddr = start_up;
+ new->size = end_down - start_up;
+ list_add_tail(&new->list, vc_list);
+ rest -= end_down - start_up;
+ }
+
+ if (rest > 0) {
+ new = get_new_element();
+ if (!new)
+ return -ENOMEM;
+ new->paddr = end_down;
+ new->size = PAGE_SIZE;
+ list_add_tail(&new->list, vc_list);
+ rest -= end - end_down;
+ }
+
+ WARN_ON(rest > 0);
/* Update the program header offset */
- phdr_ptr->p_offset = vmcore_off + (paddr - start);
+ phdr_ptr->p_offset = vmcore_off + (start - start_down);
vmcore_off = vmcore_off + size;
}
return 0;
--
1.8.3.1
--
Thanks.
HATAYAMA, Daisuke
More information about the kexec
mailing list