mirror of
https://github.com/torvalds/linux.git
synced 2026-04-18 06:44:00 -04:00
Currently the high-level xor code chunks up all operations into small units for only up to 1 + 4 vectors, and passes it to four different methods. This means the FPU/vector context is entered and left a lot for wide stripes, and a lot of indirect expensive indirect calls are performed. Switch to passing the entire gen_xor request to the low-level ops, and provide a macro to dispatch it to the existing helper. This reduce the number of indirect calls and FPU/vector context switches by a factor approaching nr_stripes / 4, and also reduces source and binary code size. Link: https://lkml.kernel.org/r/20260327061704.3707577-27-hch@lst.de Signed-off-by: Christoph Hellwig <hch@lst.de> Reviewed-by: Eric Biggers <ebiggers@kernel.org> Tested-by: Eric Biggers <ebiggers@kernel.org> Cc: Albert Ou <aou@eecs.berkeley.edu> Cc: Alexander Gordeev <agordeev@linux.ibm.com> Cc: Alexandre Ghiti <alex@ghiti.fr> Cc: Andreas Larsson <andreas@gaisler.com> Cc: Anton Ivanov <anton.ivanov@cambridgegreys.com> Cc: Ard Biesheuvel <ardb@kernel.org> Cc: Arnd Bergmann <arnd@arndb.de> Cc: "Borislav Petkov (AMD)" <bp@alien8.de> Cc: Catalin Marinas <catalin.marinas@arm.com> Cc: Chris Mason <clm@fb.com> Cc: Christian Borntraeger <borntraeger@linux.ibm.com> Cc: Dan Williams <dan.j.williams@intel.com> Cc: David S. Miller <davem@davemloft.net> Cc: David Sterba <dsterba@suse.com> Cc: Heiko Carstens <hca@linux.ibm.com> Cc: Herbert Xu <herbert@gondor.apana.org.au> Cc: "H. Peter Anvin" <hpa@zytor.com> Cc: Huacai Chen <chenhuacai@kernel.org> Cc: Ingo Molnar <mingo@redhat.com> Cc: Jason A. Donenfeld <jason@zx2c4.com> Cc: Johannes Berg <johannes@sipsolutions.net> Cc: Li Nan <linan122@huawei.com> Cc: Madhavan Srinivasan <maddy@linux.ibm.com> Cc: Magnus Lindholm <linmag7@gmail.com> Cc: Matt Turner <mattst88@gmail.com> Cc: Michael Ellerman <mpe@ellerman.id.au> Cc: Nicholas Piggin <npiggin@gmail.com> Cc: Palmer Dabbelt <palmer@dabbelt.com> Cc: Richard Henderson <richard.henderson@linaro.org> Cc: Richard Weinberger <richard@nod.at> Cc: Russell King <linux@armlinux.org.uk> Cc: Song Liu <song@kernel.org> Cc: Sven Schnelle <svens@linux.ibm.com> Cc: Ted Ts'o <tytso@mit.edu> Cc: Vasily Gorbik <gor@linux.ibm.com> Cc: WANG Xuerui <kernel@xen0n.name> Cc: Will Deacon <will@kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
134 lines
2.9 KiB
C
134 lines
2.9 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Optimized xor_block operation for RAID4/5
|
|
*
|
|
* Copyright IBM Corp. 2016
|
|
* Author(s): Martin Schwidefsky <schwidefsky@de.ibm.com>
|
|
*/
|
|
|
|
#include <linux/types.h>
|
|
#include "xor_impl.h"
|
|
#include "xor_arch.h"
|
|
|
|
static void xor_xc_2(unsigned long bytes, unsigned long * __restrict p1,
|
|
const unsigned long * __restrict p2)
|
|
{
|
|
asm volatile(
|
|
" aghi %0,-1\n"
|
|
" jm 3f\n"
|
|
" srlg 0,%0,8\n"
|
|
" ltgr 0,0\n"
|
|
" jz 1f\n"
|
|
"0: xc 0(256,%1),0(%2)\n"
|
|
" la %1,256(%1)\n"
|
|
" la %2,256(%2)\n"
|
|
" brctg 0,0b\n"
|
|
"1: exrl %0,2f\n"
|
|
" j 3f\n"
|
|
"2: xc 0(1,%1),0(%2)\n"
|
|
"3:"
|
|
: "+a" (bytes), "+a" (p1), "+a" (p2)
|
|
: : "0", "cc", "memory");
|
|
}
|
|
|
|
static void xor_xc_3(unsigned long bytes, unsigned long * __restrict p1,
|
|
const unsigned long * __restrict p2,
|
|
const unsigned long * __restrict p3)
|
|
{
|
|
asm volatile(
|
|
" aghi %0,-1\n"
|
|
" jm 4f\n"
|
|
" srlg 0,%0,8\n"
|
|
" ltgr 0,0\n"
|
|
" jz 1f\n"
|
|
"0: xc 0(256,%1),0(%2)\n"
|
|
" xc 0(256,%1),0(%3)\n"
|
|
" la %1,256(%1)\n"
|
|
" la %2,256(%2)\n"
|
|
" la %3,256(%3)\n"
|
|
" brctg 0,0b\n"
|
|
"1: exrl %0,2f\n"
|
|
" exrl %0,3f\n"
|
|
" j 4f\n"
|
|
"2: xc 0(1,%1),0(%2)\n"
|
|
"3: xc 0(1,%1),0(%3)\n"
|
|
"4:"
|
|
: "+a" (bytes), "+a" (p1), "+a" (p2), "+a" (p3)
|
|
: : "0", "cc", "memory");
|
|
}
|
|
|
|
static void xor_xc_4(unsigned long bytes, unsigned long * __restrict p1,
|
|
const unsigned long * __restrict p2,
|
|
const unsigned long * __restrict p3,
|
|
const unsigned long * __restrict p4)
|
|
{
|
|
asm volatile(
|
|
" aghi %0,-1\n"
|
|
" jm 5f\n"
|
|
" srlg 0,%0,8\n"
|
|
" ltgr 0,0\n"
|
|
" jz 1f\n"
|
|
"0: xc 0(256,%1),0(%2)\n"
|
|
" xc 0(256,%1),0(%3)\n"
|
|
" xc 0(256,%1),0(%4)\n"
|
|
" la %1,256(%1)\n"
|
|
" la %2,256(%2)\n"
|
|
" la %3,256(%3)\n"
|
|
" la %4,256(%4)\n"
|
|
" brctg 0,0b\n"
|
|
"1: exrl %0,2f\n"
|
|
" exrl %0,3f\n"
|
|
" exrl %0,4f\n"
|
|
" j 5f\n"
|
|
"2: xc 0(1,%1),0(%2)\n"
|
|
"3: xc 0(1,%1),0(%3)\n"
|
|
"4: xc 0(1,%1),0(%4)\n"
|
|
"5:"
|
|
: "+a" (bytes), "+a" (p1), "+a" (p2), "+a" (p3), "+a" (p4)
|
|
: : "0", "cc", "memory");
|
|
}
|
|
|
|
static void xor_xc_5(unsigned long bytes, unsigned long * __restrict p1,
|
|
const unsigned long * __restrict p2,
|
|
const unsigned long * __restrict p3,
|
|
const unsigned long * __restrict p4,
|
|
const unsigned long * __restrict p5)
|
|
{
|
|
asm volatile(
|
|
" aghi %0,-1\n"
|
|
" jm 6f\n"
|
|
" srlg 0,%0,8\n"
|
|
" ltgr 0,0\n"
|
|
" jz 1f\n"
|
|
"0: xc 0(256,%1),0(%2)\n"
|
|
" xc 0(256,%1),0(%3)\n"
|
|
" xc 0(256,%1),0(%4)\n"
|
|
" xc 0(256,%1),0(%5)\n"
|
|
" la %1,256(%1)\n"
|
|
" la %2,256(%2)\n"
|
|
" la %3,256(%3)\n"
|
|
" la %4,256(%4)\n"
|
|
" la %5,256(%5)\n"
|
|
" brctg 0,0b\n"
|
|
"1: exrl %0,2f\n"
|
|
" exrl %0,3f\n"
|
|
" exrl %0,4f\n"
|
|
" exrl %0,5f\n"
|
|
" j 6f\n"
|
|
"2: xc 0(1,%1),0(%2)\n"
|
|
"3: xc 0(1,%1),0(%3)\n"
|
|
"4: xc 0(1,%1),0(%4)\n"
|
|
"5: xc 0(1,%1),0(%5)\n"
|
|
"6:"
|
|
: "+a" (bytes), "+a" (p1), "+a" (p2), "+a" (p3), "+a" (p4),
|
|
"+a" (p5)
|
|
: : "0", "cc", "memory");
|
|
}
|
|
|
|
DO_XOR_BLOCKS(xc, xor_xc_2, xor_xc_3, xor_xc_4, xor_xc_5);
|
|
|
|
struct xor_block_template xor_block_xc = {
|
|
.name = "xc",
|
|
.xor_gen = xor_gen_xc,
|
|
};
|