[PATCH] arm64: Implement support for read-mostly sections
Catalin Marinas
catalin.marinas at arm.com
Tue Dec 2 03:42:48 PST 2014
On Mon, Dec 01, 2014 at 10:01:06PM +0000, Jungseok Lee wrote:
> As putting data which is read mostly together, we can avoid
> unnecessary cache line bouncing.
>
> Other architectures, such as ARM and x86, adopted the same idea.
>
> Cc: Catalin Marinas <catalin.marinas at arm.com>
> Cc: Will Deacon <will.deacon at arm.com>
> Signed-off-by: Jungseok Lee <jungseoklee85 at gmail.com>
It looks fine to me, with a nitpick below:
> ---
> arch/arm64/include/asm/cache.h | 2 ++
> 1 file changed, 2 insertions(+)
>
> diff --git a/arch/arm64/include/asm/cache.h b/arch/arm64/include/asm/cache.h
> index 88cc05b..c1a2a9f 100644
> --- a/arch/arm64/include/asm/cache.h
> +++ b/arch/arm64/include/asm/cache.h
> @@ -30,6 +30,8 @@
> */
> #define ARCH_DMA_MINALIGN L1_CACHE_BYTES
>
> +#define __read_mostly __attribute__((__section__(".data..read_mostly")))
> +
> #ifndef __ASSEMBLY__
I think we can move this #define below #ifndef as it doesn't make sense
in .S files anyway.
Acked-by: Catalin Marinas <catalin.marinas at arm.com>
More information about the linux-arm-kernel
mailing list