gcc silently ignores unsupported inline asm operand modifiers, effectively
turning '%r0' into '%0', but upcoming clang 9 complains about them:
lib/raid6/s390vx8.c:63:16: error: invalid operand in inline asm: 'VLM $2,$3,0,${1:r}'
        asm volatile ("VLM %2,%3,0,%r1"
                      ^
Clean up what look like a typo 'r' inline asm operand modifier usage.
Signed-off-by: Vasily Gorbik <gor@linux.ibm.com>
Signed-off-by: Heiko Carstens <heiko.carstens@de.ibm.com>
		
	
			
		
			
				
	
	
		
			172 lines
		
	
	
		
			3.8 KiB
		
	
	
	
		
			Ucode
		
	
	
	
	
	
			
		
		
	
	
			172 lines
		
	
	
		
			3.8 KiB
		
	
	
	
		
			Ucode
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*
 | |
|  * raid6_vx$#.c
 | |
|  *
 | |
|  * $#-way unrolled RAID6 gen/xor functions for s390
 | |
|  * based on the vector facility
 | |
|  *
 | |
|  * Copyright IBM Corp. 2016
 | |
|  * Author(s): Martin Schwidefsky <schwidefsky@de.ibm.com>
 | |
|  *
 | |
|  * This file is postprocessed using unroll.awk.
 | |
|  */
 | |
| 
 | |
| #include <linux/raid/pq.h>
 | |
| #include <asm/fpu/api.h>
 | |
| 
 | |
| asm(".include \"asm/vx-insn.h\"\n");
 | |
| 
 | |
| #define NSIZE 16
 | |
| 
 | |
| static inline void LOAD_CONST(void)
 | |
| {
 | |
| 	asm volatile("VREPIB %v24,7");
 | |
| 	asm volatile("VREPIB %v25,0x1d");
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * The SHLBYTE() operation shifts each of the 16 bytes in
 | |
|  * vector register y left by 1 bit and stores the result in
 | |
|  * vector register x.
 | |
|  */
 | |
| static inline void SHLBYTE(int x, int y)
 | |
| {
 | |
| 	asm volatile ("VAB %0,%1,%1" : : "i" (x), "i" (y));
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * For each of the 16 bytes in the vector register y the MASK()
 | |
|  * operation returns 0xFF if the high bit of the byte is 1,
 | |
|  * or 0x00 if the high bit is 0. The result is stored in vector
 | |
|  * register x.
 | |
|  */
 | |
| static inline void MASK(int x, int y)
 | |
| {
 | |
| 	asm volatile ("VESRAVB	%0,%1,24" : : "i" (x), "i" (y));
 | |
| }
 | |
| 
 | |
| static inline void AND(int x, int y, int z)
 | |
| {
 | |
| 	asm volatile ("VN %0,%1,%2" : : "i" (x), "i" (y), "i" (z));
 | |
| }
 | |
| 
 | |
| static inline void XOR(int x, int y, int z)
 | |
| {
 | |
| 	asm volatile ("VX %0,%1,%2" : : "i" (x), "i" (y), "i" (z));
 | |
| }
 | |
| 
 | |
| static inline void LOAD_DATA(int x, u8 *ptr)
 | |
| {
 | |
| 	typedef struct { u8 _[16 * $#]; } addrtype;
 | |
| 	register addrtype *__ptr asm("1") = (addrtype *) ptr;
 | |
| 
 | |
| 	asm volatile ("VLM %2,%3,0,%1"
 | |
| 		      : : "m" (*__ptr), "a" (__ptr), "i" (x),
 | |
| 			  "i" (x + $# - 1));
 | |
| }
 | |
| 
 | |
| static inline void STORE_DATA(int x, u8 *ptr)
 | |
| {
 | |
| 	typedef struct { u8 _[16 * $#]; } addrtype;
 | |
| 	register addrtype *__ptr asm("1") = (addrtype *) ptr;
 | |
| 
 | |
| 	asm volatile ("VSTM %2,%3,0,1"
 | |
| 		      : "=m" (*__ptr) : "a" (__ptr), "i" (x),
 | |
| 			"i" (x + $# - 1));
 | |
| }
 | |
| 
 | |
| static inline void COPY_VEC(int x, int y)
 | |
| {
 | |
| 	asm volatile ("VLR %0,%1" : : "i" (x), "i" (y));
 | |
| }
 | |
| 
 | |
| static void raid6_s390vx$#_gen_syndrome(int disks, size_t bytes, void **ptrs)
 | |
| {
 | |
| 	struct kernel_fpu vxstate;
 | |
| 	u8 **dptr, *p, *q;
 | |
| 	int d, z, z0;
 | |
| 
 | |
| 	kernel_fpu_begin(&vxstate, KERNEL_VXR);
 | |
| 	LOAD_CONST();
 | |
| 
 | |
| 	dptr = (u8 **) ptrs;
 | |
| 	z0 = disks - 3;		/* Highest data disk */
 | |
| 	p = dptr[z0 + 1];	/* XOR parity */
 | |
| 	q = dptr[z0 + 2];	/* RS syndrome */
 | |
| 
 | |
| 	for (d = 0; d < bytes; d += $#*NSIZE) {
 | |
| 		LOAD_DATA(0,&dptr[z0][d]);
 | |
| 		COPY_VEC(8+$$,0+$$);
 | |
| 		for (z = z0 - 1; z >= 0; z--) {
 | |
| 			MASK(16+$$,8+$$);
 | |
| 			AND(16+$$,16+$$,25);
 | |
| 			SHLBYTE(8+$$,8+$$);
 | |
| 			XOR(8+$$,8+$$,16+$$);
 | |
| 			LOAD_DATA(16,&dptr[z][d]);
 | |
| 			XOR(0+$$,0+$$,16+$$);
 | |
| 			XOR(8+$$,8+$$,16+$$);
 | |
| 		}
 | |
| 		STORE_DATA(0,&p[d]);
 | |
| 		STORE_DATA(8,&q[d]);
 | |
| 	}
 | |
| 	kernel_fpu_end(&vxstate, KERNEL_VXR);
 | |
| }
 | |
| 
 | |
| static void raid6_s390vx$#_xor_syndrome(int disks, int start, int stop,
 | |
| 					size_t bytes, void **ptrs)
 | |
| {
 | |
| 	struct kernel_fpu vxstate;
 | |
| 	u8 **dptr, *p, *q;
 | |
| 	int d, z, z0;
 | |
| 
 | |
| 	dptr = (u8 **) ptrs;
 | |
| 	z0 = stop;		/* P/Q right side optimization */
 | |
| 	p = dptr[disks - 2];	/* XOR parity */
 | |
| 	q = dptr[disks - 1];	/* RS syndrome */
 | |
| 
 | |
| 	kernel_fpu_begin(&vxstate, KERNEL_VXR);
 | |
| 	LOAD_CONST();
 | |
| 
 | |
| 	for (d = 0; d < bytes; d += $#*NSIZE) {
 | |
| 		/* P/Q data pages */
 | |
| 		LOAD_DATA(0,&dptr[z0][d]);
 | |
| 		COPY_VEC(8+$$,0+$$);
 | |
| 		for (z = z0 - 1; z >= start; z--) {
 | |
| 			MASK(16+$$,8+$$);
 | |
| 			AND(16+$$,16+$$,25);
 | |
| 			SHLBYTE(8+$$,8+$$);
 | |
| 			XOR(8+$$,8+$$,16+$$);
 | |
| 			LOAD_DATA(16,&dptr[z][d]);
 | |
| 			XOR(0+$$,0+$$,16+$$);
 | |
| 			XOR(8+$$,8+$$,16+$$);
 | |
| 		}
 | |
| 		/* P/Q left side optimization */
 | |
| 		for (z = start - 1; z >= 0; z--) {
 | |
| 			MASK(16+$$,8+$$);
 | |
| 			AND(16+$$,16+$$,25);
 | |
| 			SHLBYTE(8+$$,8+$$);
 | |
| 			XOR(8+$$,8+$$,16+$$);
 | |
| 		}
 | |
| 		LOAD_DATA(16,&p[d]);
 | |
| 		XOR(16+$$,16+$$,0+$$);
 | |
| 		STORE_DATA(16,&p[d]);
 | |
| 		LOAD_DATA(16,&q[d]);
 | |
| 		XOR(16+$$,16+$$,8+$$);
 | |
| 		STORE_DATA(16,&q[d]);
 | |
| 	}
 | |
| 	kernel_fpu_end(&vxstate, KERNEL_VXR);
 | |
| }
 | |
| 
 | |
| static int raid6_s390vx$#_valid(void)
 | |
| {
 | |
| 	return MACHINE_HAS_VX;
 | |
| }
 | |
| 
 | |
| const struct raid6_calls raid6_s390vx$# = {
 | |
| 	raid6_s390vx$#_gen_syndrome,
 | |
| 	raid6_s390vx$#_xor_syndrome,
 | |
| 	raid6_s390vx$#_valid,
 | |
| 	"vx128x$#",
 | |
| 	1
 | |
| };
 |