[PATCH v5 4/8] vmalloc: introduce remap_vmalloc_range_partial

Vivek Goyal vgoyal at redhat.com
Tue May 14 10:59:06 EDT 2013


On Tue, May 14, 2013 at 10:57:29AM +0900, HATAYAMA Daisuke wrote:
> We want to allocate ELF note segment buffer on the 2nd kernel in
> vmalloc space and remap it to user-space in order to reduce the risk
> that memory allocation fails on system with huge number of CPUs and so
> with huge ELF note segment that exceeds 11-order block size.
> 
> Although there's already remap_vmalloc_range for the purpose of
> remapping vmalloc memory to user-space, we need to specify user-space
> range via vma. Mmap on /proc/vmcore needs to remap range across
> multiple objects, so the interface that requires vma to cover full
> range is problematic.
> 
> This patch introduces remap_vmalloc_range_partial that receives
> user-space range as a pair of base address and size and can be used
> for mmap on /proc/vmcore case.
> 
> remap_vmalloc_range is rewritten using remap_vmalloc_range_partial.
> 
> Signed-off-by: HATAYAMA Daisuke <d.hatayama at jp.fujitsu.com>

This also needs ACK of VM folks. CCing some of them.

Thanks
Vivek

> ---
> 
>  include/linux/vmalloc.h |    4 +++
>  mm/vmalloc.c            |   63 +++++++++++++++++++++++++++++++++--------------
>  2 files changed, 48 insertions(+), 19 deletions(-)
> 
> diff --git a/include/linux/vmalloc.h b/include/linux/vmalloc.h
> index 7d5773a..dd0a2c8 100644
> --- a/include/linux/vmalloc.h
> +++ b/include/linux/vmalloc.h
> @@ -82,6 +82,10 @@ extern void *vmap(struct page **pages, unsigned int count,
>  			unsigned long flags, pgprot_t prot);
>  extern void vunmap(const void *addr);
>  
> +extern int remap_vmalloc_range_partial(struct vm_area_struct *vma,
> +				       unsigned long uaddr, void *kaddr,
> +				       unsigned long size);
> +

>  extern int remap_vmalloc_range(struct vm_area_struct *vma, void *addr,
>  							unsigned long pgoff);
>  void vmalloc_sync_all(void);
> diff --git a/mm/vmalloc.c b/mm/vmalloc.c
> index 3875fa2..d9a9f4f6 100644
> --- a/mm/vmalloc.c
> +++ b/mm/vmalloc.c
> @@ -2148,42 +2148,44 @@ finished:
>  }
>  
>  /**
> - *	remap_vmalloc_range  -  map vmalloc pages to userspace
> - *	@vma:		vma to cover (map full range of vma)
> - *	@addr:		vmalloc memory
> - *	@pgoff:		number of pages into addr before first page to map
> + *	remap_vmalloc_range_partial  -  map vmalloc pages to userspace
> + *	@vma:		vma to cover
> + *	@uaddr:		target user address to start at
> + *	@kaddr:		virtual address of vmalloc kernel memory
> + *	@size:		size of map area
>   *
>   *	Returns:	0 for success, -Exxx on failure
>   *
> - *	This function checks that addr is a valid vmalloc'ed area, and
> - *	that it is big enough to cover the vma. Will return failure if
> - *	that criteria isn't met.
> + *	This function checks that @kaddr is a valid vmalloc'ed area,
> + *	and that it is big enough to cover the range starting at
> + *	@uaddr in @vma. Will return failure if that criteria isn't
> + *	met.
>   *
>   *	Similar to remap_pfn_range() (see mm/memory.c)
>   */
> -int remap_vmalloc_range(struct vm_area_struct *vma, void *addr,
> -						unsigned long pgoff)
> +int remap_vmalloc_range_partial(struct vm_area_struct *vma, unsigned long uaddr,
> +				void *kaddr, unsigned long size)
>  {
>  	struct vm_struct *area;
> -	unsigned long uaddr = vma->vm_start;
> -	unsigned long usize = vma->vm_end - vma->vm_start;
>  
> -	if ((PAGE_SIZE-1) & (unsigned long)addr)
> +	size = PAGE_ALIGN(size);
> +
> +	if (((PAGE_SIZE-1) & (unsigned long)uaddr) ||
> +	    ((PAGE_SIZE-1) & (unsigned long)kaddr))
>  		return -EINVAL;
>  
> -	area = find_vm_area(addr);
> +	area = find_vm_area(kaddr);
>  	if (!area)
>  		return -EINVAL;
>  
>  	if (!(area->flags & VM_USERMAP))
>  		return -EINVAL;
>  
> -	if (usize + (pgoff << PAGE_SHIFT) > area->size - PAGE_SIZE)
> +	if (kaddr + size > area->addr + area->size)
>  		return -EINVAL;
>  
> -	addr += pgoff << PAGE_SHIFT;
>  	do {
> -		struct page *page = vmalloc_to_page(addr);
> +		struct page *page = vmalloc_to_page(kaddr);
>  		int ret;
>  
>  		ret = vm_insert_page(vma, uaddr, page);
> @@ -2191,14 +2193,37 @@ int remap_vmalloc_range(struct vm_area_struct *vma, void *addr,
>  			return ret;
>  
>  		uaddr += PAGE_SIZE;
> -		addr += PAGE_SIZE;
> -		usize -= PAGE_SIZE;
> -	} while (usize > 0);
> +		kaddr += PAGE_SIZE;
> +		size -= PAGE_SIZE;
> +	} while (size > 0);
>  
>  	vma->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP;
>  
>  	return 0;
>  }
> +EXPORT_SYMBOL(remap_vmalloc_range_partial);
> +
> +/**
> + *	remap_vmalloc_range  -  map vmalloc pages to userspace
> + *	@vma:		vma to cover (map full range of vma)
> + *	@addr:		vmalloc memory
> + *	@pgoff:		number of pages into addr before first page to map
> + *
> + *	Returns:	0 for success, -Exxx on failure
> + *
> + *	This function checks that addr is a valid vmalloc'ed area, and
> + *	that it is big enough to cover the vma. Will return failure if
> + *	that criteria isn't met.
> + *
> + *	Similar to remap_pfn_range() (see mm/memory.c)
> + */
> +int remap_vmalloc_range(struct vm_area_struct *vma, void *addr,
> +						unsigned long pgoff)
> +{
> +	return remap_vmalloc_range_partial(vma, vma->vm_start,
> +					   addr + (pgoff << PAGE_SHIFT),
> +					   vma->vm_end - vma->vm_start);
> +}
>  EXPORT_SYMBOL(remap_vmalloc_range);
>  
>  /*



More information about the kexec mailing list