mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	Writes to ZCR_EL1 are self-synchronising, and so may be expensive in typical implementations. This patch adopts the approach used for costly system register writes elsewhere in the kernel: the system register write is suppressed if it would not change the stored value. Since the common case will be that of switching between tasks that use the same vector length as one another, prediction hit rates on the conditional branch should be reasonably good, with lower expected amortised cost than the unconditional execution of a heavyweight self-synchronising instruction. Signed-off-by: Dave Martin <Dave.Martin@arm.com> Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
		
			
				
	
	
		
			227 lines
		
	
	
	
		
			5.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			227 lines
		
	
	
	
		
			5.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * FP/SIMD state saving and restoring macros
 | 
						|
 *
 | 
						|
 * Copyright (C) 2012 ARM Ltd.
 | 
						|
 * Author: Catalin Marinas <catalin.marinas@arm.com>
 | 
						|
 *
 | 
						|
 * This program is free software; you can redistribute it and/or modify
 | 
						|
 * it under the terms of the GNU General Public License version 2 as
 | 
						|
 * published by the Free Software Foundation.
 | 
						|
 *
 | 
						|
 * This program is distributed in the hope that it will be useful,
 | 
						|
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
						|
 * GNU General Public License for more details.
 | 
						|
 *
 | 
						|
 * You should have received a copy of the GNU General Public License
 | 
						|
 * along with this program.  If not, see <http://www.gnu.org/licenses/>.
 | 
						|
 */
 | 
						|
 | 
						|
.macro fpsimd_save state, tmpnr
 | 
						|
	stp	q0, q1, [\state, #16 * 0]
 | 
						|
	stp	q2, q3, [\state, #16 * 2]
 | 
						|
	stp	q4, q5, [\state, #16 * 4]
 | 
						|
	stp	q6, q7, [\state, #16 * 6]
 | 
						|
	stp	q8, q9, [\state, #16 * 8]
 | 
						|
	stp	q10, q11, [\state, #16 * 10]
 | 
						|
	stp	q12, q13, [\state, #16 * 12]
 | 
						|
	stp	q14, q15, [\state, #16 * 14]
 | 
						|
	stp	q16, q17, [\state, #16 * 16]
 | 
						|
	stp	q18, q19, [\state, #16 * 18]
 | 
						|
	stp	q20, q21, [\state, #16 * 20]
 | 
						|
	stp	q22, q23, [\state, #16 * 22]
 | 
						|
	stp	q24, q25, [\state, #16 * 24]
 | 
						|
	stp	q26, q27, [\state, #16 * 26]
 | 
						|
	stp	q28, q29, [\state, #16 * 28]
 | 
						|
	stp	q30, q31, [\state, #16 * 30]!
 | 
						|
	mrs	x\tmpnr, fpsr
 | 
						|
	str	w\tmpnr, [\state, #16 * 2]
 | 
						|
	mrs	x\tmpnr, fpcr
 | 
						|
	str	w\tmpnr, [\state, #16 * 2 + 4]
 | 
						|
.endm
 | 
						|
 | 
						|
.macro fpsimd_restore_fpcr state, tmp
 | 
						|
	/*
 | 
						|
	 * Writes to fpcr may be self-synchronising, so avoid restoring
 | 
						|
	 * the register if it hasn't changed.
 | 
						|
	 */
 | 
						|
	mrs	\tmp, fpcr
 | 
						|
	cmp	\tmp, \state
 | 
						|
	b.eq	9999f
 | 
						|
	msr	fpcr, \state
 | 
						|
9999:
 | 
						|
.endm
 | 
						|
 | 
						|
/* Clobbers \state */
 | 
						|
.macro fpsimd_restore state, tmpnr
 | 
						|
	ldp	q0, q1, [\state, #16 * 0]
 | 
						|
	ldp	q2, q3, [\state, #16 * 2]
 | 
						|
	ldp	q4, q5, [\state, #16 * 4]
 | 
						|
	ldp	q6, q7, [\state, #16 * 6]
 | 
						|
	ldp	q8, q9, [\state, #16 * 8]
 | 
						|
	ldp	q10, q11, [\state, #16 * 10]
 | 
						|
	ldp	q12, q13, [\state, #16 * 12]
 | 
						|
	ldp	q14, q15, [\state, #16 * 14]
 | 
						|
	ldp	q16, q17, [\state, #16 * 16]
 | 
						|
	ldp	q18, q19, [\state, #16 * 18]
 | 
						|
	ldp	q20, q21, [\state, #16 * 20]
 | 
						|
	ldp	q22, q23, [\state, #16 * 22]
 | 
						|
	ldp	q24, q25, [\state, #16 * 24]
 | 
						|
	ldp	q26, q27, [\state, #16 * 26]
 | 
						|
	ldp	q28, q29, [\state, #16 * 28]
 | 
						|
	ldp	q30, q31, [\state, #16 * 30]!
 | 
						|
	ldr	w\tmpnr, [\state, #16 * 2]
 | 
						|
	msr	fpsr, x\tmpnr
 | 
						|
	ldr	w\tmpnr, [\state, #16 * 2 + 4]
 | 
						|
	fpsimd_restore_fpcr x\tmpnr, \state
 | 
						|
.endm
 | 
						|
 | 
						|
/* Sanity-check macros to help avoid encoding garbage instructions */
 | 
						|
 | 
						|
.macro _check_general_reg nr
 | 
						|
	.if (\nr) < 0 || (\nr) > 30
 | 
						|
		.error "Bad register number \nr."
 | 
						|
	.endif
 | 
						|
.endm
 | 
						|
 | 
						|
.macro _sve_check_zreg znr
 | 
						|
	.if (\znr) < 0 || (\znr) > 31
 | 
						|
		.error "Bad Scalable Vector Extension vector register number \znr."
 | 
						|
	.endif
 | 
						|
.endm
 | 
						|
 | 
						|
.macro _sve_check_preg pnr
 | 
						|
	.if (\pnr) < 0 || (\pnr) > 15
 | 
						|
		.error "Bad Scalable Vector Extension predicate register number \pnr."
 | 
						|
	.endif
 | 
						|
.endm
 | 
						|
 | 
						|
.macro _check_num n, min, max
 | 
						|
	.if (\n) < (\min) || (\n) > (\max)
 | 
						|
		.error "Number \n out of range [\min,\max]"
 | 
						|
	.endif
 | 
						|
.endm
 | 
						|
 | 
						|
/* SVE instruction encodings for non-SVE-capable assemblers */
 | 
						|
 | 
						|
/* STR (vector): STR Z\nz, [X\nxbase, #\offset, MUL VL] */
 | 
						|
.macro _sve_str_v nz, nxbase, offset=0
 | 
						|
	_sve_check_zreg \nz
 | 
						|
	_check_general_reg \nxbase
 | 
						|
	_check_num (\offset), -0x100, 0xff
 | 
						|
	.inst	0xe5804000			\
 | 
						|
		| (\nz)				\
 | 
						|
		| ((\nxbase) << 5)		\
 | 
						|
		| (((\offset) & 7) << 10)	\
 | 
						|
		| (((\offset) & 0x1f8) << 13)
 | 
						|
.endm
 | 
						|
 | 
						|
/* LDR (vector): LDR Z\nz, [X\nxbase, #\offset, MUL VL] */
 | 
						|
.macro _sve_ldr_v nz, nxbase, offset=0
 | 
						|
	_sve_check_zreg \nz
 | 
						|
	_check_general_reg \nxbase
 | 
						|
	_check_num (\offset), -0x100, 0xff
 | 
						|
	.inst	0x85804000			\
 | 
						|
		| (\nz)				\
 | 
						|
		| ((\nxbase) << 5)		\
 | 
						|
		| (((\offset) & 7) << 10)	\
 | 
						|
		| (((\offset) & 0x1f8) << 13)
 | 
						|
.endm
 | 
						|
 | 
						|
/* STR (predicate): STR P\np, [X\nxbase, #\offset, MUL VL] */
 | 
						|
.macro _sve_str_p np, nxbase, offset=0
 | 
						|
	_sve_check_preg \np
 | 
						|
	_check_general_reg \nxbase
 | 
						|
	_check_num (\offset), -0x100, 0xff
 | 
						|
	.inst	0xe5800000			\
 | 
						|
		| (\np)				\
 | 
						|
		| ((\nxbase) << 5)		\
 | 
						|
		| (((\offset) & 7) << 10)	\
 | 
						|
		| (((\offset) & 0x1f8) << 13)
 | 
						|
.endm
 | 
						|
 | 
						|
/* LDR (predicate): LDR P\np, [X\nxbase, #\offset, MUL VL] */
 | 
						|
.macro _sve_ldr_p np, nxbase, offset=0
 | 
						|
	_sve_check_preg \np
 | 
						|
	_check_general_reg \nxbase
 | 
						|
	_check_num (\offset), -0x100, 0xff
 | 
						|
	.inst	0x85800000			\
 | 
						|
		| (\np)				\
 | 
						|
		| ((\nxbase) << 5)		\
 | 
						|
		| (((\offset) & 7) << 10)	\
 | 
						|
		| (((\offset) & 0x1f8) << 13)
 | 
						|
.endm
 | 
						|
 | 
						|
/* RDVL X\nx, #\imm */
 | 
						|
.macro _sve_rdvl nx, imm
 | 
						|
	_check_general_reg \nx
 | 
						|
	_check_num (\imm), -0x20, 0x1f
 | 
						|
	.inst	0x04bf5000			\
 | 
						|
		| (\nx)				\
 | 
						|
		| (((\imm) & 0x3f) << 5)
 | 
						|
.endm
 | 
						|
 | 
						|
/* RDFFR (unpredicated): RDFFR P\np.B */
 | 
						|
.macro _sve_rdffr np
 | 
						|
	_sve_check_preg \np
 | 
						|
	.inst	0x2519f000			\
 | 
						|
		| (\np)
 | 
						|
.endm
 | 
						|
 | 
						|
/* WRFFR P\np.B */
 | 
						|
.macro _sve_wrffr np
 | 
						|
	_sve_check_preg \np
 | 
						|
	.inst	0x25289000			\
 | 
						|
		| ((\np) << 5)
 | 
						|
.endm
 | 
						|
 | 
						|
.macro __for from:req, to:req
 | 
						|
	.if (\from) == (\to)
 | 
						|
		_for__body \from
 | 
						|
	.else
 | 
						|
		__for \from, (\from) + ((\to) - (\from)) / 2
 | 
						|
		__for (\from) + ((\to) - (\from)) / 2 + 1, \to
 | 
						|
	.endif
 | 
						|
.endm
 | 
						|
 | 
						|
.macro _for var:req, from:req, to:req, insn:vararg
 | 
						|
	.macro _for__body \var:req
 | 
						|
		\insn
 | 
						|
	.endm
 | 
						|
 | 
						|
	__for \from, \to
 | 
						|
 | 
						|
	.purgem _for__body
 | 
						|
.endm
 | 
						|
 | 
						|
.macro sve_save nxbase, xpfpsr, nxtmp
 | 
						|
 _for n, 0, 31,	_sve_str_v	\n, \nxbase, \n - 34
 | 
						|
 _for n, 0, 15,	_sve_str_p	\n, \nxbase, \n - 16
 | 
						|
		_sve_rdffr	0
 | 
						|
		_sve_str_p	0, \nxbase
 | 
						|
		_sve_ldr_p	0, \nxbase, -16
 | 
						|
 | 
						|
		mrs		x\nxtmp, fpsr
 | 
						|
		str		w\nxtmp, [\xpfpsr]
 | 
						|
		mrs		x\nxtmp, fpcr
 | 
						|
		str		w\nxtmp, [\xpfpsr, #4]
 | 
						|
.endm
 | 
						|
 | 
						|
.macro sve_load nxbase, xpfpsr, xvqminus1, nxtmp, xtmp2
 | 
						|
		mrs_s		x\nxtmp, SYS_ZCR_EL1
 | 
						|
		bic		\xtmp2, x\nxtmp, ZCR_ELx_LEN_MASK
 | 
						|
		orr		\xtmp2, \xtmp2, \xvqminus1
 | 
						|
		cmp		\xtmp2, x\nxtmp
 | 
						|
		b.eq		921f
 | 
						|
		msr_s		SYS_ZCR_EL1, \xtmp2	// self-synchronising
 | 
						|
921:
 | 
						|
 _for n, 0, 31,	_sve_ldr_v	\n, \nxbase, \n - 34
 | 
						|
		_sve_ldr_p	0, \nxbase
 | 
						|
		_sve_wrffr	0
 | 
						|
 _for n, 0, 15,	_sve_ldr_p	\n, \nxbase, \n - 16
 | 
						|
 | 
						|
		ldr		w\nxtmp, [\xpfpsr]
 | 
						|
		msr		fpsr, x\nxtmp
 | 
						|
		ldr		w\nxtmp, [\xpfpsr, #4]
 | 
						|
		msr		fpcr, x\nxtmp
 | 
						|
.endm
 |