[PATCH] arm64: Implement support for read-mostly sections

Catalin Marinas catalin.marinas at arm.com
Tue Dec 2 03:42:48 PST 2014


On Mon, Dec 01, 2014 at 10:01:06PM +0000, Jungseok Lee wrote:
> As putting data which is read mostly together, we can avoid
> unnecessary cache line bouncing.
> 
> Other architectures, such as ARM and x86, adopted the same idea.
> 
> Cc: Catalin Marinas <catalin.marinas at arm.com>
> Cc: Will Deacon <will.deacon at arm.com>
> Signed-off-by: Jungseok Lee <jungseoklee85 at gmail.com>

It looks fine to me, with a nitpick below:

> ---
>  arch/arm64/include/asm/cache.h | 2 ++
>  1 file changed, 2 insertions(+)
> 
> diff --git a/arch/arm64/include/asm/cache.h b/arch/arm64/include/asm/cache.h
> index 88cc05b..c1a2a9f 100644
> --- a/arch/arm64/include/asm/cache.h
> +++ b/arch/arm64/include/asm/cache.h
> @@ -30,6 +30,8 @@
>   */
>  #define ARCH_DMA_MINALIGN	L1_CACHE_BYTES
>  
> +#define __read_mostly __attribute__((__section__(".data..read_mostly")))
> +
>  #ifndef __ASSEMBLY__

I think we can move this #define below #ifndef as it doesn't make sense
in .S files anyway.

Acked-by: Catalin Marinas <catalin.marinas at arm.com>



More information about the linux-arm-kernel mailing list