[PATCH 4/9] swap: Simplify shmem_unuse() usage [optional]

Hugh Dickins hugh at veritas.com
Fri Mar 2 11:44:38 EST 2007


On Fri, 2 Mar 2007, Richard Purdie wrote:

> Simplify shmem_unuse_inode() removing a confusing optimisation which
> requires the caller to call swap_duplicate if the shmem_unuse() call
> doesn't succeed.
> 
> Based on a patch by Nick Piggin and some of my own changes as discussed
> on LKML. 
> 
> Signed-off-by: Richard Purdie <rpurdie at openedhand.com>

Definite NAK to this one from me: I'm sorry the optimization confuses
you, but it's well commented at both ends, and speeds up shmem swapoff
very significantly e.g. minutes down to seconds.  There may well be a
less confusing way of achieving the same effect, with another return
code from shmem_unuse, and some gotos, but I'm not all that keen.

Your other patches, well, as ever I hope I'll get to look at them,
but there are so many people, all much quicker than me, playing in
mm these days...

Hugh

> 
> ---
>  mm/shmem.c    |   12 +++++-------
>  mm/swapfile.c |   23 ++---------------------
>  2 files changed, 7 insertions(+), 28 deletions(-)
> 
> Index: linux/mm/shmem.c
> ===================================================================
> --- linux.orig/mm/shmem.c	2007-02-28 18:12:34.000000000 +0000
> +++ linux/mm/shmem.c	2007-02-28 18:12:46.000000000 +0000
> @@ -734,7 +734,7 @@ static int shmem_unuse_inode(struct shme
>  	struct page **dir;
>  	struct page *subdir;
>  	swp_entry_t *ptr;
> -	int offset;
> +	int offset, moved;
>  
>  	idx = 0;
>  	ptr = info->i_direct;
> @@ -792,17 +792,15 @@ lost2:
>  found:
>  	idx += offset;
>  	inode = &info->vfs_inode;
> -	if (move_from_swap_cache(page, idx, inode->i_mapping) == 0) {
> +	moved = (move_from_swap_cache(page, idx, inode->i_mapping) == 0);
> +	if (moved) {
>  		info->flags |= SHMEM_PAGEIN;
>  		shmem_swp_set(info, ptr + offset, 0);
>  	}
>  	shmem_swp_unmap(ptr);
>  	spin_unlock(&info->lock);
> -	/*
> -	 * Decrement swap count even when the entry is left behind:
> -	 * try_to_unuse will skip over mms, then reincrement count.
> -	 */
> -	swap_free(entry, page);
> +	if (moved)
> +		swap_free(entry, page);
>  	return 1;
>  }
>  
> Index: linux/mm/swapfile.c
> ===================================================================
> --- linux.orig/mm/swapfile.c	2007-02-28 18:12:41.000000000 +0000
> +++ linux/mm/swapfile.c	2007-02-28 18:13:04.000000000 +0000
> @@ -689,15 +689,6 @@ void try_to_unuse_page_entry(struct page
>  	if (!shmem_unuse(entry, page)) {
>  		try_to_unuse_anon(entry, page);
>  		delete_from_swap_cache(page);
> -	} else if (PageSwapCache(page)) {
> -		/*
> -		 * shmem_unuse deleted a swappage from the swap cache, but the
> -		 * move to filepage failed so it left swappage in cache and
> -		 * lowered its swap count to pass quickly through the loops in
> -		 * try_to_unuse(). We must reincrement the count to try again
> -		 * later (ick).
> -		 */
> -		swap_duplicate(entry);
>  	}
>  }
>  
> @@ -922,12 +913,6 @@ static int try_to_unuse(unsigned int typ
>  		 * read from disk into another page.  Splitting into two
>  		 * pages would be incorrect if swap supported "shared
>  		 * private" pages, but they are handled by tmpfs files.
> -		 *
> -		 * Note shmem_unuse already deleted a swappage from
> -		 * the swap cache, unless the move to filepage failed:
> -		 * in which case it left swappage in cache, lowered its
> -		 * swap count to pass quickly through the loops above,
> -		 * and now we must reincrement count to try again later.
>  		 */
>  		if ((*swap_map > 1) && PageDirty(page) && PageSwapCache(page)) {
>  			struct writeback_control wbc = {
> @@ -938,12 +923,8 @@ static int try_to_unuse(unsigned int typ
>  			lock_page(page);
>  			wait_on_page_writeback(page);
>  		}
> -		if (PageSwapCache(page)) {
> -			if (shmem)
> -				swap_duplicate(entry);
> -			else
> -				delete_from_swap_cache(page);
> -		}
> +		if (PageSwapCache(page) && !shmem)
> +			delete_from_swap_cache(page);
>  
>  		/*
>  		 * So we could skip searching mms once swap count went




More information about the linux-mtd mailing list