[PATCH 1/1] arm64: Add ARM64 optimized IP checksum routine

Robin Murphy robin.murphy at arm.com
Wed May 11 05:55:20 PDT 2016


Hi Luke,

On 10/05/16 19:27, Luke Starrett wrote:
> This change implements an optimized checksum for arm64, based loosely on
> the original arch/arm implementation.  Load-pair is used for the initial
> 16B load, reducing the overall number of loads to two for packets without
> IP options.  Instruction count is reduced by ~3x compared to generic C
> implementation.  Changes were tested against LE and BE operation and for
> all possible mis-alignments.
>
> Signed-off-by: Luke Starrett <luke.starrett at broadcom.com>
> ---
>   arch/arm64/include/asm/checksum.h | 57 +++++++++++++++++++++++++++++++++++++++
>   1 file changed, 57 insertions(+)
>   create mode 100644 arch/arm64/include/asm/checksum.h
>
> diff --git a/arch/arm64/include/asm/checksum.h b/arch/arm64/include/asm/checksum.h
> new file mode 100644
> index 0000000..15629d7
> --- /dev/null
> +++ b/arch/arm64/include/asm/checksum.h
> @@ -0,0 +1,57 @@
> +/*
> + * Copyright 2016 Broadcom
> + *
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License, version 2, as
> + * published by the Free Software Foundation (the "GPL").
> + *
> + * This program is distributed in the hope that it will be useful, but
> + * WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> + * General Public License version 2 (GPLv2) for more details.
> + *
> + * You should have received a copy of the GNU General Public License
> + * version 2 (GPLv2) along with this source code.
> + *
> + * ip_fast_csum() loosely derived from original arch/arm implementation
> + */
> +
> +#ifndef __ASM_ARM64_CHECKSUM_H
> +#define __ASM_ARM64_CHECKSUM_H
> +
> +/*
> + * This is a version of ip_compute_csum() optimized for IP headers,
> + * which always checksum on 4 octet boundaries.
> + */
> +static inline __sum16
> +ip_fast_csum(const void *iph, unsigned int ihl)
> +{
> +	u64 tmp, sum;
> +
> +	__asm__ __volatile__ (
> +"	ldp	%0, %3, [%1], #16\n"
> +"	sub	%2, %2, #4\n"
> +"	adds	%0, %0, %3\n"
> +"1:	ldr	%w3, [%1], #4\n"
> +"	sub	%2, %2, #1\n"
> +"	adcs	%0, %0, %3\n"

This looks suspicious - the following tst is going to zero the carry 
flag, so either setting the flags is unnecessary, or things are working 
by chance.

> +"	tst	%2, #15\n"
> +"	bne	1b\n"
> +"	adc	%0, %0, xzr\n"

Similarly, you should never get here with the carry flag set, so this is 
just a pointless nop.

> +"	ror	%3, %0, #32\n"
> +"	add     %0, %0, %3\n"
> +"	lsr	%0, %0, #32\n"
> +"	ror	%w3, %w0, #16\n"
> +"	add     %w0, %w0, %w3\n"
> +"	lsr	%0, %0, #16\n"
> +	: "=r" (sum), "=r" (iph), "=r" (ihl), "=r" (tmp)
> +	: "1" (iph), "2" (ihl)
> +	: "cc", "memory");
> +
> +	return (__force __sum16)(~sum);
> +}

Given that dodginess, and that there doesn't seem to be anything here 
which absolutely _demands_ assembly voodoo, can we not just have an 
optimised C implementation working on larger chunks, along the lines of 
Alpha?

Robin.

> +
> +#define ip_fast_csum ip_fast_csum
> +#include <asm-generic/checksum.h>
> +
> +#endif	/* __ASM_ARM64_CHECKSUM_H */
>




More information about the linux-arm-kernel mailing list